var/home/core/zuul-output/0000755000175000017500000000000015067245733014541 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067256374015510 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004642451015067256365017722 0ustar rootrootOct 01 15:46:49 crc systemd[1]: Starting Kubernetes Kubelet... Oct 01 15:46:49 crc restorecon[4587]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:49 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 15:46:50 crc restorecon[4587]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 01 15:46:51 crc kubenswrapper[4688]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.092091 4688 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.098983 4688 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099029 4688 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099044 4688 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099056 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099067 4688 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099108 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099123 4688 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099139 4688 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099150 4688 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099161 4688 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099175 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099187 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099197 4688 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099208 4688 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099219 4688 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099229 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099239 4688 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099250 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099259 4688 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099268 4688 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099277 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099288 4688 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099298 4688 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099308 4688 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099318 4688 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099328 4688 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099338 4688 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099349 4688 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099360 4688 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099371 4688 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099381 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099393 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099406 4688 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099418 4688 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099432 4688 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099444 4688 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099458 4688 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099468 4688 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099479 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099489 4688 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099501 4688 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099515 4688 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099565 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099582 4688 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099595 4688 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099606 4688 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099617 4688 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099628 4688 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099642 4688 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099655 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099667 4688 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099678 4688 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099690 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099701 4688 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099712 4688 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099727 4688 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099741 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099752 4688 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099763 4688 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099774 4688 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099784 4688 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099794 4688 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099805 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099819 4688 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099834 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099844 4688 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099854 4688 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099863 4688 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099873 4688 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099885 4688 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.099895 4688 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100124 4688 flags.go:64] FLAG: --address="0.0.0.0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100157 4688 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100181 4688 flags.go:64] FLAG: --anonymous-auth="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100196 4688 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100214 4688 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100226 4688 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100243 4688 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100257 4688 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100270 4688 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100283 4688 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100296 4688 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100309 4688 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100320 4688 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100332 4688 flags.go:64] FLAG: --cgroup-root="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100344 4688 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100355 4688 flags.go:64] FLAG: --client-ca-file="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100366 4688 flags.go:64] FLAG: --cloud-config="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100377 4688 flags.go:64] FLAG: --cloud-provider="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100388 4688 flags.go:64] FLAG: --cluster-dns="[]" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100403 4688 flags.go:64] FLAG: --cluster-domain="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100415 4688 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100428 4688 flags.go:64] FLAG: --config-dir="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100440 4688 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100452 4688 flags.go:64] FLAG: --container-log-max-files="5" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100469 4688 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100481 4688 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100494 4688 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100507 4688 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100555 4688 flags.go:64] FLAG: --contention-profiling="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100570 4688 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100583 4688 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100596 4688 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100608 4688 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100624 4688 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100636 4688 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100650 4688 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100661 4688 flags.go:64] FLAG: --enable-load-reader="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100677 4688 flags.go:64] FLAG: --enable-server="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100689 4688 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100706 4688 flags.go:64] FLAG: --event-burst="100" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100718 4688 flags.go:64] FLAG: --event-qps="50" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100730 4688 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100742 4688 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100753 4688 flags.go:64] FLAG: --eviction-hard="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100769 4688 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100781 4688 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100793 4688 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100805 4688 flags.go:64] FLAG: --eviction-soft="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100817 4688 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100827 4688 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100839 4688 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100851 4688 flags.go:64] FLAG: --experimental-mounter-path="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100862 4688 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100873 4688 flags.go:64] FLAG: --fail-swap-on="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100885 4688 flags.go:64] FLAG: --feature-gates="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100899 4688 flags.go:64] FLAG: --file-check-frequency="20s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100913 4688 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100926 4688 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100939 4688 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100951 4688 flags.go:64] FLAG: --healthz-port="10248" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100963 4688 flags.go:64] FLAG: --help="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100976 4688 flags.go:64] FLAG: --hostname-override="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.100988 4688 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101000 4688 flags.go:64] FLAG: --http-check-frequency="20s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101012 4688 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101023 4688 flags.go:64] FLAG: --image-credential-provider-config="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101036 4688 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101048 4688 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101060 4688 flags.go:64] FLAG: --image-service-endpoint="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101073 4688 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101084 4688 flags.go:64] FLAG: --kube-api-burst="100" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101096 4688 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101108 4688 flags.go:64] FLAG: --kube-api-qps="50" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101123 4688 flags.go:64] FLAG: --kube-reserved="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101135 4688 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101146 4688 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101158 4688 flags.go:64] FLAG: --kubelet-cgroups="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101169 4688 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101181 4688 flags.go:64] FLAG: --lock-file="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101192 4688 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101204 4688 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101216 4688 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101235 4688 flags.go:64] FLAG: --log-json-split-stream="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101246 4688 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101257 4688 flags.go:64] FLAG: --log-text-split-stream="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101269 4688 flags.go:64] FLAG: --logging-format="text" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101280 4688 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101302 4688 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101314 4688 flags.go:64] FLAG: --manifest-url="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101326 4688 flags.go:64] FLAG: --manifest-url-header="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101342 4688 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101355 4688 flags.go:64] FLAG: --max-open-files="1000000" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101370 4688 flags.go:64] FLAG: --max-pods="110" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101382 4688 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101393 4688 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101405 4688 flags.go:64] FLAG: --memory-manager-policy="None" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101416 4688 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101428 4688 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101440 4688 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101452 4688 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101483 4688 flags.go:64] FLAG: --node-status-max-images="50" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101494 4688 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101506 4688 flags.go:64] FLAG: --oom-score-adj="-999" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101518 4688 flags.go:64] FLAG: --pod-cidr="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101563 4688 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101583 4688 flags.go:64] FLAG: --pod-manifest-path="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101594 4688 flags.go:64] FLAG: --pod-max-pids="-1" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101606 4688 flags.go:64] FLAG: --pods-per-core="0" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101618 4688 flags.go:64] FLAG: --port="10250" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101630 4688 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101643 4688 flags.go:64] FLAG: --provider-id="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101655 4688 flags.go:64] FLAG: --qos-reserved="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101667 4688 flags.go:64] FLAG: --read-only-port="10255" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101680 4688 flags.go:64] FLAG: --register-node="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101692 4688 flags.go:64] FLAG: --register-schedulable="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101703 4688 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101723 4688 flags.go:64] FLAG: --registry-burst="10" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101734 4688 flags.go:64] FLAG: --registry-qps="5" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101746 4688 flags.go:64] FLAG: --reserved-cpus="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101765 4688 flags.go:64] FLAG: --reserved-memory="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101781 4688 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101793 4688 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101805 4688 flags.go:64] FLAG: --rotate-certificates="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101817 4688 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101829 4688 flags.go:64] FLAG: --runonce="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101840 4688 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101851 4688 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101863 4688 flags.go:64] FLAG: --seccomp-default="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101877 4688 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101889 4688 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101901 4688 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101913 4688 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101925 4688 flags.go:64] FLAG: --storage-driver-password="root" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101936 4688 flags.go:64] FLAG: --storage-driver-secure="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101949 4688 flags.go:64] FLAG: --storage-driver-table="stats" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101960 4688 flags.go:64] FLAG: --storage-driver-user="root" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101971 4688 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101983 4688 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.101994 4688 flags.go:64] FLAG: --system-cgroups="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102006 4688 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102025 4688 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102036 4688 flags.go:64] FLAG: --tls-cert-file="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102047 4688 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102061 4688 flags.go:64] FLAG: --tls-min-version="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102073 4688 flags.go:64] FLAG: --tls-private-key-file="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102087 4688 flags.go:64] FLAG: --topology-manager-policy="none" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102098 4688 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102109 4688 flags.go:64] FLAG: --topology-manager-scope="container" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102122 4688 flags.go:64] FLAG: --v="2" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102137 4688 flags.go:64] FLAG: --version="false" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102152 4688 flags.go:64] FLAG: --vmodule="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102172 4688 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.102185 4688 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103199 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103221 4688 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103230 4688 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103238 4688 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103246 4688 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103254 4688 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103262 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103273 4688 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103283 4688 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103293 4688 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103304 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103313 4688 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103322 4688 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103331 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103341 4688 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103351 4688 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103360 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103369 4688 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103377 4688 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103385 4688 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103393 4688 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103401 4688 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103409 4688 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103417 4688 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103426 4688 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103434 4688 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103443 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103452 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103462 4688 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103470 4688 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103481 4688 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103490 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103498 4688 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103506 4688 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103515 4688 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103553 4688 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103561 4688 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103569 4688 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103577 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103585 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103593 4688 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103601 4688 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103610 4688 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103620 4688 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103628 4688 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103636 4688 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103643 4688 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103651 4688 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103659 4688 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103667 4688 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103675 4688 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103683 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103690 4688 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103698 4688 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103707 4688 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103714 4688 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103722 4688 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103730 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103738 4688 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103747 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103758 4688 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103769 4688 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103780 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103790 4688 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103801 4688 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103812 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103825 4688 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103837 4688 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103847 4688 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103858 4688 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.103868 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.103900 4688 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.119850 4688 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.119903 4688 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120023 4688 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120037 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120072 4688 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120082 4688 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120092 4688 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120105 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120116 4688 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120130 4688 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120145 4688 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120156 4688 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120165 4688 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120175 4688 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120184 4688 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120193 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120202 4688 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120212 4688 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120221 4688 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120233 4688 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120242 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120252 4688 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120261 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120272 4688 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120281 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120290 4688 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120298 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120307 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120316 4688 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120325 4688 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120333 4688 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120341 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120349 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120359 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120367 4688 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120379 4688 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120390 4688 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120400 4688 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120410 4688 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120419 4688 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120427 4688 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120436 4688 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120444 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120453 4688 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120461 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120470 4688 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120478 4688 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120489 4688 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120499 4688 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120508 4688 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120518 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120552 4688 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120562 4688 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120571 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120579 4688 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120587 4688 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120595 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120603 4688 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120611 4688 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120619 4688 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120627 4688 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120634 4688 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120642 4688 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120651 4688 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120658 4688 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120666 4688 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120674 4688 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120682 4688 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120689 4688 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120697 4688 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120704 4688 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120712 4688 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120721 4688 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.120734 4688 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120953 4688 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120965 4688 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120974 4688 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120982 4688 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120990 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.120998 4688 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121008 4688 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121017 4688 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121025 4688 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121032 4688 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121040 4688 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121048 4688 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121056 4688 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121063 4688 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121071 4688 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121078 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121086 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121096 4688 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121106 4688 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121117 4688 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121128 4688 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121141 4688 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121152 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121162 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121172 4688 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121180 4688 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121188 4688 feature_gate.go:330] unrecognized feature gate: Example Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121196 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121203 4688 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121211 4688 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121219 4688 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121227 4688 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121234 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121242 4688 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121250 4688 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121257 4688 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121266 4688 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121277 4688 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121288 4688 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121298 4688 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121308 4688 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121317 4688 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121325 4688 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121334 4688 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121341 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121351 4688 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121361 4688 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121372 4688 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121386 4688 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121401 4688 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121414 4688 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121424 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121436 4688 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121446 4688 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121454 4688 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121463 4688 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121471 4688 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121479 4688 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121487 4688 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121495 4688 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121502 4688 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121510 4688 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121518 4688 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121552 4688 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121560 4688 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121568 4688 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121576 4688 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121583 4688 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121591 4688 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121600 4688 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.121609 4688 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.121621 4688 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.122722 4688 server.go:940] "Client rotation is on, will bootstrap in background" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.130903 4688 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.131073 4688 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.133453 4688 server.go:997] "Starting client certificate rotation" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.133510 4688 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.135275 4688 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 04:05:04.252502901 +0000 UTC Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.135379 4688 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 948h18m13.117127054s for next certificate rotation Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.166104 4688 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.171687 4688 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.196616 4688 log.go:25] "Validated CRI v1 runtime API" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.235468 4688 log.go:25] "Validated CRI v1 image API" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.237735 4688 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.248284 4688 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-01-15-23-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.248323 4688 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.262249 4688 manager.go:217] Machine: {Timestamp:2025-10-01 15:46:51.260660594 +0000 UTC m=+0.611300596 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:09adb3c3-2d03-448f-a3d6-41cc884a6e83 BootID:71eff274-2d68-4a0d-9181-6c8c42ab6723 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:9d:04:ac Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:9d:04:ac Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:45:00:a9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:79:8a:d4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:42:2c:a7 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:cd:0f:15 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:16:69:a1:46:3b:6d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3e:69:88:88:12:0f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.262629 4688 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.262823 4688 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.263148 4688 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.263370 4688 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.263413 4688 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.265026 4688 topology_manager.go:138] "Creating topology manager with none policy" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.265052 4688 container_manager_linux.go:303] "Creating device plugin manager" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.265585 4688 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.265648 4688 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.266008 4688 state_mem.go:36] "Initialized new in-memory state store" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.266177 4688 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.270899 4688 kubelet.go:418] "Attempting to sync node with API server" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.270944 4688 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.270972 4688 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.270994 4688 kubelet.go:324] "Adding apiserver pod source" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.271017 4688 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.275407 4688 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.276836 4688 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.279650 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.279793 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.280290 4688 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.280346 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.280485 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282154 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282202 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282227 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282246 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282277 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282295 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.282309 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.283000 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.283060 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.283086 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.285929 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.285972 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.286022 4688 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.287024 4688 server.go:1280] "Started kubelet" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.287952 4688 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.287726 4688 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.288850 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.289086 4688 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 01 15:46:51 crc systemd[1]: Started Kubernetes Kubelet. Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.297663 4688 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.298059 4688 server.go:460] "Adding debug handlers to kubelet server" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.298583 4688 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.299774 4688 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.300195 4688 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.300366 4688 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.301041 4688 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.298656 4688 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 12:38:23.76936039 +0000 UTC Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.301893 4688 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 956h51m32.467504122s for next certificate rotation Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.301974 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.302093 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.303220 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="200ms" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.303660 4688 factory.go:55] Registering systemd factory Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.303685 4688 factory.go:221] Registration of the systemd container factory successfully Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304093 4688 factory.go:153] Registering CRI-O factory Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304108 4688 factory.go:221] Registration of the crio container factory successfully Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304183 4688 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304225 4688 factory.go:103] Registering Raw factory Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304242 4688 manager.go:1196] Started watching for new ooms in manager Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.304879 4688 manager.go:319] Starting recovery of all containers Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.303270 4688 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a68991e2a3a78 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 15:46:51.286968952 +0000 UTC m=+0.637608964,LastTimestamp:2025-10-01 15:46:51.286968952 +0000 UTC m=+0.637608964,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325095 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325160 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325173 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325185 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325194 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325204 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325216 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325227 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325241 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.325250 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327016 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327059 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327076 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327101 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327119 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327129 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327140 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327155 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327168 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327184 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327279 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327292 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327311 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327323 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327335 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327352 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327373 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327388 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327404 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327417 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327433 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327444 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327454 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327467 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327483 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327500 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327511 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327534 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327548 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327560 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327571 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327585 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327596 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327610 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327621 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327631 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327645 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327657 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327672 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327683 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327694 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327710 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327732 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327744 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327761 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327771 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327786 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327798 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327808 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327820 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327835 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327845 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327857 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327873 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327884 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.327897 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331349 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331373 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331385 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331396 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331418 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331434 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331452 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331467 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331478 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331495 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331505 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331532 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331546 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331558 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331575 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331591 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331611 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331626 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331637 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331652 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331665 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331679 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331690 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331701 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331720 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331732 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331751 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331765 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331779 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331798 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331810 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331844 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331864 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331876 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331897 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331910 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.331924 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.336873 4688 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.337015 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.337295 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339166 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339228 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339252 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339292 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339314 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339330 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339374 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.339446 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340583 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340610 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340626 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340642 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340658 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340672 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340688 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340703 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340717 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340731 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340745 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340760 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340774 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340789 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340805 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340818 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340832 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340845 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340861 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340892 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340906 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340925 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340941 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340955 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340970 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.340986 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341001 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341016 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341058 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341075 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341090 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341105 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341245 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341267 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341281 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341297 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341312 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341326 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341341 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341356 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341373 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341388 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341406 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341420 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341433 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341449 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341463 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341479 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341495 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341511 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341543 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341560 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341573 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341588 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341603 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341619 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341634 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341650 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341664 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341680 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341696 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341710 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341725 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341739 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341755 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341770 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341784 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341800 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341815 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341830 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341845 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341860 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341875 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341889 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341903 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341918 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341934 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341957 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341974 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.341992 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342005 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342019 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342034 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342050 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342065 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342080 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342095 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342109 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342125 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342140 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342154 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342170 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342186 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342201 4688 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342215 4688 reconstruct.go:97] "Volume reconstruction finished" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.342226 4688 reconciler.go:26] "Reconciler: start to sync state" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.348648 4688 manager.go:324] Recovery completed Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.360437 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.362096 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.362161 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.362177 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.363036 4688 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.363058 4688 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.363116 4688 state_mem.go:36] "Initialized new in-memory state store" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.375409 4688 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.376877 4688 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.379598 4688 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.379651 4688 kubelet.go:2335] "Starting kubelet main sync loop" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.379837 4688 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.380859 4688 policy_none.go:49] "None policy: Start" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.381329 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.381396 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.382163 4688 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.382199 4688 state_mem.go:35] "Initializing new in-memory state store" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.401735 4688 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.426800 4688 manager.go:334] "Starting Device Plugin manager" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.426982 4688 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.426997 4688 server.go:79] "Starting device plugin registration server" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.427419 4688 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.427432 4688 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.427809 4688 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.427919 4688 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.427928 4688 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.433906 4688 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.479918 4688 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.480001 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.480876 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.480918 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.480928 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.481072 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.481256 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.481301 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482029 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482068 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482167 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482180 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482210 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482428 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482464 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482706 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482739 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482826 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482925 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.482950 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483600 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483625 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483662 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483682 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483695 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483816 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.483976 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.484001 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485717 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485750 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485763 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485893 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.485987 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.486023 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488124 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488167 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488186 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488282 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488305 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.488322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.504262 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="400ms" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.528306 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.529348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.529370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.529379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.529399 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.529992 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543597 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543634 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543654 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543669 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543686 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543701 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543716 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543736 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543752 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543792 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543842 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543867 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543890 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543909 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.543930 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.644847 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.644909 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645002 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645032 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645207 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645295 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645368 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645395 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645438 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645407 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645489 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645503 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645518 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645553 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645588 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645644 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645602 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645644 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645766 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645832 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645900 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.645974 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646022 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646059 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646054 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646166 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646123 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646236 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646269 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.646311 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.731018 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.733486 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.733551 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.733562 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.733586 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.734090 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.833149 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.859925 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.875584 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.881442 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-63086eacf957ec54f222205dc5e8d5a02588a5be540f0d072d92f6724e9193bf WatchSource:0}: Error finding container 63086eacf957ec54f222205dc5e8d5a02588a5be540f0d072d92f6724e9193bf: Status 404 returned error can't find the container with id 63086eacf957ec54f222205dc5e8d5a02588a5be540f0d072d92f6724e9193bf Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.893685 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7cf07a233b05dea2a54eb677f1793489580b3d3f09a90249b1c6305fd90ccd44 WatchSource:0}: Error finding container 7cf07a233b05dea2a54eb677f1793489580b3d3f09a90249b1c6305fd90ccd44: Status 404 returned error can't find the container with id 7cf07a233b05dea2a54eb677f1793489580b3d3f09a90249b1c6305fd90ccd44 Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.894280 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.895029 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-df1aa9f37bfd07725b846f75da19cb9b604c2095c1546ea7296d6af6c4f4c1fd WatchSource:0}: Error finding container df1aa9f37bfd07725b846f75da19cb9b604c2095c1546ea7296d6af6c4f4c1fd: Status 404 returned error can't find the container with id df1aa9f37bfd07725b846f75da19cb9b604c2095c1546ea7296d6af6c4f4c1fd Oct 01 15:46:51 crc kubenswrapper[4688]: I1001 15:46:51.903619 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:51 crc kubenswrapper[4688]: E1001 15:46:51.904890 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="800ms" Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.925270 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2823e1d5d77682d5886f55e7d8bddf35c4ff1cb76e894dd5f52b23102c71edd0 WatchSource:0}: Error finding container 2823e1d5d77682d5886f55e7d8bddf35c4ff1cb76e894dd5f52b23102c71edd0: Status 404 returned error can't find the container with id 2823e1d5d77682d5886f55e7d8bddf35c4ff1cb76e894dd5f52b23102c71edd0 Oct 01 15:46:51 crc kubenswrapper[4688]: W1001 15:46:51.927464 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-876589efdae6dfd7d2b9cdf7883e2f7dcca584454975001f6981a074f1e69705 WatchSource:0}: Error finding container 876589efdae6dfd7d2b9cdf7883e2f7dcca584454975001f6981a074f1e69705: Status 404 returned error can't find the container with id 876589efdae6dfd7d2b9cdf7883e2f7dcca584454975001f6981a074f1e69705 Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.134344 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.135664 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.135707 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.135716 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.135738 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.136120 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.290640 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:52 crc kubenswrapper[4688]: W1001 15:46:52.337551 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.337635 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.386599 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"876589efdae6dfd7d2b9cdf7883e2f7dcca584454975001f6981a074f1e69705"} Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.388634 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df1aa9f37bfd07725b846f75da19cb9b604c2095c1546ea7296d6af6c4f4c1fd"} Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.389514 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7cf07a233b05dea2a54eb677f1793489580b3d3f09a90249b1c6305fd90ccd44"} Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.390139 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"63086eacf957ec54f222205dc5e8d5a02588a5be540f0d072d92f6724e9193bf"} Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.390939 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2823e1d5d77682d5886f55e7d8bddf35c4ff1cb76e894dd5f52b23102c71edd0"} Oct 01 15:46:52 crc kubenswrapper[4688]: W1001 15:46:52.434327 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.434403 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.706762 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="1.6s" Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.735725 4688 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a68991e2a3a78 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 15:46:51.286968952 +0000 UTC m=+0.637608964,LastTimestamp:2025-10-01 15:46:51.286968952 +0000 UTC m=+0.637608964,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 15:46:52 crc kubenswrapper[4688]: W1001 15:46:52.772616 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.772875 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:52 crc kubenswrapper[4688]: W1001 15:46:52.791216 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.791412 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.936613 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.938322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.938401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.938444 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:52 crc kubenswrapper[4688]: I1001 15:46:52.938484 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:52 crc kubenswrapper[4688]: E1001 15:46:52.939282 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.290005 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.397074 4688 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd" exitCode=0 Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.397220 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.397258 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.398677 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.398731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.398754 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.399519 4688 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d" exitCode=0 Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.399675 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.399735 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.401239 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.401273 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.401292 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.403155 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.403942 4688 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3d8d2e399fd984f443ac17d8a8b5b414ae2e80c99253669aeadb647968de2352" exitCode=0 Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.403968 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404013 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404017 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404141 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3d8d2e399fd984f443ac17d8a8b5b414ae2e80c99253669aeadb647968de2352"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404182 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404841 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404870 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.404885 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.406607 4688 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3" exitCode=0 Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.406718 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.406889 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.408158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.408185 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.408198 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.411609 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.411668 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25"} Oct 01 15:46:53 crc kubenswrapper[4688]: I1001 15:46:53.411680 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0"} Oct 01 15:46:54 crc kubenswrapper[4688]: W1001 15:46:54.282934 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:54 crc kubenswrapper[4688]: E1001 15:46:54.283001 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.289684 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:54 crc kubenswrapper[4688]: E1001 15:46:54.313960 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="3.2s" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.417662 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.417762 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.418805 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.418835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.418848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.420682 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.420741 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.420764 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.420768 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.421793 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.421831 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.421840 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.423594 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.423633 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.423649 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.425910 4688 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="194b6619ec16c3f4d974978e756766107bcd194a9521d6cba3edb5a1007338d7" exitCode=0 Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.425998 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"194b6619ec16c3f4d974978e756766107bcd194a9521d6cba3edb5a1007338d7"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.426057 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.427006 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.427032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.427041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.428612 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.428629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d2931483443bdbbcf2d60d3d9e7f8e616a7a61c8ba9a05eb7d0b9534a4c3ba90"} Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.429418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.429451 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.429466 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.539693 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.543073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.543100 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.543108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.543129 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:54 crc kubenswrapper[4688]: E1001 15:46:54.543408 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 01 15:46:54 crc kubenswrapper[4688]: I1001 15:46:54.704468 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:54 crc kubenswrapper[4688]: W1001 15:46:54.957234 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:54 crc kubenswrapper[4688]: E1001 15:46:54.957352 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.289460 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:55 crc kubenswrapper[4688]: W1001 15:46:55.298106 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:55 crc kubenswrapper[4688]: E1001 15:46:55.298194 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.435467 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45e421840a3ff25e9ed20ff4f9b61565aeb91c478abf9613398df44c267c3c57"} Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.435513 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.435623 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9"} Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.436380 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.436414 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.436425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.437747 4688 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="09cf3e7b7acebf3a3d1f445453f7c37a66b90825a19f7ec36427ef5efab7e9ae" exitCode=0 Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.437791 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"09cf3e7b7acebf3a3d1f445453f7c37a66b90825a19f7ec36427ef5efab7e9ae"} Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.437837 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.437897 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.437947 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438045 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438078 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438592 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438619 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438632 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438642 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438678 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438765 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438781 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438930 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438957 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:55 crc kubenswrapper[4688]: I1001 15:46:55.438971 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:55 crc kubenswrapper[4688]: W1001 15:46:55.445687 4688 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:55 crc kubenswrapper[4688]: E1001 15:46:55.445858 4688 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.290833 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444551 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c0afd840a5f390ee0c4495c9c29727bb8c1eea52989dde047e5b5997bcefe742"} Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444600 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9a208c76f4df936e8318484c8d8c675709650b7336846d8c682bd916cc42aae2"} Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444644 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444698 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444701 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.444649 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446594 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446634 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446654 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446666 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446693 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446705 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446778 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.446800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.840024 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.852787 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:56 crc kubenswrapper[4688]: I1001 15:46:56.945680 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.451923 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.458940 4688 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45e421840a3ff25e9ed20ff4f9b61565aeb91c478abf9613398df44c267c3c57" exitCode=255 Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.459040 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"45e421840a3ff25e9ed20ff4f9b61565aeb91c478abf9613398df44c267c3c57"} Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.459257 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.460265 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.460315 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.460331 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.461085 4688 scope.go:117] "RemoveContainer" containerID="45e421840a3ff25e9ed20ff4f9b61565aeb91c478abf9613398df44c267c3c57" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.466406 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4131fe16a6db17f0af91280da094a901bf88adf9f0fcdf4a49273a36058528cf"} Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.466460 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c3d0c7ae5f0c60429d296180372de3b8b67d521980298ca7b11a7e25f2c8701a"} Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.466478 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9eceeea346b530f926dd97dbc9f2b40ed918d1855bea16c4d279e658a8ade5e"} Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.466491 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.466491 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.467773 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.467777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.467824 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.467938 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.469699 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.469749 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.704470 4688 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.704580 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.743749 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.744915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.744943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.744951 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:57 crc kubenswrapper[4688]: I1001 15:46:57.744969 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.272560 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.473761 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.476114 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721"} Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.476302 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.476419 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.476498 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.476834 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478179 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478242 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478456 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478494 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.478507 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.479339 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.479411 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.479431 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:58 crc kubenswrapper[4688]: I1001 15:46:58.597574 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.024157 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.478978 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.479151 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.479187 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.479228 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480085 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480097 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480830 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480891 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480927 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480961 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.480989 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:46:59 crc kubenswrapper[4688]: I1001 15:46:59.597563 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:00 crc kubenswrapper[4688]: I1001 15:47:00.052353 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:00 crc kubenswrapper[4688]: I1001 15:47:00.481103 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:00 crc kubenswrapper[4688]: I1001 15:47:00.482373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:00 crc kubenswrapper[4688]: I1001 15:47:00.482440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:00 crc kubenswrapper[4688]: I1001 15:47:00.482460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.394973 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.395225 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.396737 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.396807 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.396837 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:01 crc kubenswrapper[4688]: E1001 15:47:01.434221 4688 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.483798 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.485345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.485409 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:01 crc kubenswrapper[4688]: I1001 15:47:01.485430 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:06 crc kubenswrapper[4688]: I1001 15:47:06.950077 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:47:06 crc kubenswrapper[4688]: I1001 15:47:06.950198 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:06 crc kubenswrapper[4688]: I1001 15:47:06.951038 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:06 crc kubenswrapper[4688]: I1001 15:47:06.951067 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:06 crc kubenswrapper[4688]: I1001 15:47:06.951079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.291159 4688 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.339556 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.339668 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.347056 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.347155 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.704996 4688 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:47:07 crc kubenswrapper[4688]: I1001 15:47:07.705078 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:47:09 crc kubenswrapper[4688]: I1001 15:47:09.598248 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 15:47:09 crc kubenswrapper[4688]: I1001 15:47:09.598342 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.058697 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.058933 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.059701 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.059811 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.060184 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.060248 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.060276 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.065860 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.505250 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.505515 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.505574 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.506423 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.506447 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:10 crc kubenswrapper[4688]: I1001 15:47:10.506455 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.420881 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.421127 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.422298 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.422335 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.422344 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:11 crc kubenswrapper[4688]: E1001 15:47:11.434314 4688 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.439088 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.508016 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.509070 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.509111 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.509126 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.971303 4688 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 15:47:11 crc kubenswrapper[4688]: I1001 15:47:11.971394 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 15:47:12 crc kubenswrapper[4688]: E1001 15:47:12.321267 4688 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.324957 4688 trace.go:236] Trace[1642969616]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:47:00.061) (total time: 12262ms): Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[1642969616]: ---"Objects listed" error: 12262ms (15:47:12.324) Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[1642969616]: [12.262954458s] [12.262954458s] END Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.325244 4688 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.326811 4688 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.326946 4688 trace.go:236] Trace[780336783]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:46:59.931) (total time: 12394ms): Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[780336783]: ---"Objects listed" error: 12394ms (15:47:12.326) Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[780336783]: [12.3949498s] [12.3949498s] END Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.326982 4688 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 01 15:47:12 crc kubenswrapper[4688]: E1001 15:47:12.326898 4688 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.328001 4688 trace.go:236] Trace[4465219]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:46:57.749) (total time: 14578ms): Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[4465219]: ---"Objects listed" error: 14578ms (15:47:12.327) Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[4465219]: [14.578504078s] [14.578504078s] END Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.328045 4688 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.336458 4688 trace.go:236] Trace[1303392243]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 15:46:58.811) (total time: 13524ms): Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[1303392243]: ---"Objects listed" error: 13524ms (15:47:12.336) Oct 01 15:47:12 crc kubenswrapper[4688]: Trace[1303392243]: [13.524850386s] [13.524850386s] END Oct 01 15:47:12 crc kubenswrapper[4688]: I1001 15:47:12.336785 4688 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.284444 4688 apiserver.go:52] "Watching apiserver" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.287866 4688 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288129 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-xgf9s","openshift-machine-config-operator/machine-config-daemon-6ndlj","openshift-multus/multus-jcfs4","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-xq8gs","openshift-multus/multus-additional-cni-plugins-nrzjb"] Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288435 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.288492 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288644 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288727 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288797 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288867 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.288859 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.288911 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.288974 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.289573 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.289720 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.289830 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.289859 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.289610 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.293118 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.293423 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.293511 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.293703 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.293983 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.294140 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.294430 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.294517 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.294680 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.298232 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.298462 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.298474 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.301701 4688 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.302803 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.302871 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.302803 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303133 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303252 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303474 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303640 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303725 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303939 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304033 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304057 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.303678 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304144 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304233 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304264 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304291 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304454 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304489 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.304540 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.319485 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331859 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331889 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331904 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331926 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331953 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331977 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.331996 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332015 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332040 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332080 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332100 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332118 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332152 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332175 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332193 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332216 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332231 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332249 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332272 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332287 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332301 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332316 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332330 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332346 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332363 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332379 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332397 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332414 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332431 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332464 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332479 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332494 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332510 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332595 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332614 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.332810 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333158 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333292 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333377 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333401 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333419 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333434 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333451 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333471 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333494 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333508 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333546 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333562 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333577 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333593 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333609 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333624 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333638 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333653 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333668 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333682 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333702 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333718 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333734 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333750 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333766 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333781 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333795 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333819 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333833 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333847 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333862 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333878 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333894 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333910 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333925 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333940 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.333958 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334028 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334063 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334082 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334098 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334116 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334136 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334154 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334177 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334200 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334219 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334237 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334256 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334279 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334300 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334320 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334343 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334364 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334386 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334409 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334426 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334445 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334463 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334481 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334497 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334515 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334546 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334571 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334591 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334611 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334630 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334650 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334670 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334690 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334710 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334734 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334751 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334768 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334786 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334810 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334834 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334853 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334868 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334884 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334902 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334916 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334932 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334949 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334965 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334981 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.334998 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335015 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335030 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335046 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335062 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335078 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335093 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335175 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335195 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335215 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335231 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335248 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335266 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335281 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335298 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335315 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335330 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335349 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335365 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335385 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335401 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335417 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335433 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335449 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335466 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335483 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335499 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335515 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335553 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335567 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335582 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335598 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335613 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335631 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335647 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335663 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335679 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335696 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335713 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335729 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335745 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335762 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335778 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335794 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335812 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335828 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335844 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335861 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335880 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335899 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335915 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335931 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335948 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335964 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335981 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.335997 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336013 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336030 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336046 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336063 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336080 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336097 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336114 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336132 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336147 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336165 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336182 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336199 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336215 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336231 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336247 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336263 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336279 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336296 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336312 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336328 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336376 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69fd7067-ac62-426d-b3a3-dca2750d14db-rootfs\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336394 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336409 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336432 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-system-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336483 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-netns\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336498 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336515 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336884 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336903 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-socket-dir-parent\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336922 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336941 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336956 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-multus\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336971 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.336990 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337007 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337025 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6900ee15-3789-4e35-8d10-573acac2f5bd-hosts-file\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337042 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-etc-kubernetes\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337057 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337075 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337090 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-daemon-config\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337110 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337137 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337154 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7cn\" (UniqueName: \"kubernetes.io/projected/69fd7067-ac62-426d-b3a3-dca2750d14db-kube-api-access-lq7cn\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337170 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-system-cni-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337187 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-conf-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337205 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337222 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337238 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8bsc\" (UniqueName: \"kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337253 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337268 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-k8s-cni-cncf-io\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337283 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337297 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337312 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-os-release\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337327 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-kubelet\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337343 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337359 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337384 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337399 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-hostroot\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337416 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2p6x\" (UniqueName: \"kubernetes.io/projected/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-kube-api-access-j2p6x\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337435 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337450 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cnibin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337484 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337500 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337517 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337549 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cnibin\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337566 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337582 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cni-binary-copy\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-multus-certs\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337617 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337633 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337648 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337666 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337681 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69fd7067-ac62-426d-b3a3-dca2750d14db-proxy-tls\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337697 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69fd7067-ac62-426d-b3a3-dca2750d14db-mcd-auth-proxy-config\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337713 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-os-release\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337729 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-bin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337746 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wrpp\" (UniqueName: \"kubernetes.io/projected/e83f8b9c-424d-4983-b5d3-83b52939c5eb-kube-api-access-6wrpp\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337761 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337777 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337796 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337811 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337829 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88k4h\" (UniqueName: \"kubernetes.io/projected/6900ee15-3789-4e35-8d10-573acac2f5bd-kube-api-access-88k4h\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337872 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.337884 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.341863 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.342017 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.342157 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.343517 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.343748 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.344188 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.344419 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.344574 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:13.844524228 +0000 UTC m=+23.195164220 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.345573 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.346109 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.346598 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.347112 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.347252 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.347377 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.347501 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.349133 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.349354 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.351787 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.352344 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.352371 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.352438 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:13.852424943 +0000 UTC m=+23.203064905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.352848 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.353962 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.354246 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.354933 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.357557 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.357935 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.358020 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.358319 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.358587 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.359046 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.359060 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.359513 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.359759 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.361905 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.362159 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.362163 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.362701 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.365567 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.365735 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.365864 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366008 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366036 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366059 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366117 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366896 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366372 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366508 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366597 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366734 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366838 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.366168 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367006 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367119 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367245 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367392 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367423 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367476 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367582 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367704 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367741 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367744 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367800 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367869 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.367897 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.368369 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.368688 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.368983 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.368887 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369200 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369288 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369299 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369552 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369683 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.369992 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.370036 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.370229 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.370264 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.346095 4688 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.370374 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.370640 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.371251 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.371474 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.371611 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.371827 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.371845 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.372046 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.372315 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.372334 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.372517 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.372903 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.373167 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.373415 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.373605 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.373609 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.373846 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.374029 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.374285 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.374328 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:13.874305597 +0000 UTC m=+23.224945559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376650 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376573 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.374577 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.374945 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376737 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375164 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375179 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375182 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375269 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375349 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375483 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376825 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375501 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375702 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.375846 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376129 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.376850 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377061 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377120 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377154 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377239 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377246 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377361 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377601 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377793 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.377857 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.378024 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.378314 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.378384 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.378425 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.378409 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379063 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379255 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379242 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379556 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379644 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.379984 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380012 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380071 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380278 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.380357 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.383272 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.383284 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.383332 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:13.883318143 +0000 UTC m=+23.233958105 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380742 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380303 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380423 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.383163 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.380999 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.381537 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.381569 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.381629 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.381679 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.382195 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.382344 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.382592 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.382827 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.383071 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.383157 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.383755 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.384031 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.386092 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388135 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.387127 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.387291 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.387507 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388641 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388735 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388522 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388082 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388126 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388808 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.388946 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.387847 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389001 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389021 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389169 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389412 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389623 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389737 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.389936 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.390410 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.390497 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.390578 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.390868 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.390927 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.391024 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.391649 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.391716 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.391858 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392113 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392299 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392326 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392697 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392729 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.392772 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.393941 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.394027 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.394618 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.394664 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.394703 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.395103 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.395763 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.396649 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.398494 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.398514 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.398672 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.398717 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:13.898703572 +0000 UTC m=+23.249343534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.398972 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.401713 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.402856 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.403605 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.404539 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.405331 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.406117 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.406644 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.403702 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.403728 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.403740 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.403866 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.418147 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.419072 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.419167 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.419583 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.430945 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.432492 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439087 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69fd7067-ac62-426d-b3a3-dca2750d14db-rootfs\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439283 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439486 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439592 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439689 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-system-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439773 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-netns\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439862 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.439947 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440022 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-socket-dir-parent\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440119 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-multus\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440200 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440279 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-etc-kubernetes\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440435 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6900ee15-3789-4e35-8d10-573acac2f5bd-hosts-file\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440542 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-daemon-config\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440619 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-conf-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440832 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7cn\" (UniqueName: \"kubernetes.io/projected/69fd7067-ac62-426d-b3a3-dca2750d14db-kube-api-access-lq7cn\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.440926 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-system-cni-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441009 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-k8s-cni-cncf-io\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441141 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/69fd7067-ac62-426d-b3a3-dca2750d14db-rootfs\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441248 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441632 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441760 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441847 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8bsc\" (UniqueName: \"kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441932 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442021 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442101 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442186 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442259 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-os-release\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442334 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-kubelet\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442424 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2p6x\" (UniqueName: \"kubernetes.io/projected/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-kube-api-access-j2p6x\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442512 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442598 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-hostroot\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442668 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442745 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442831 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cnibin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442903 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442985 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443072 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cnibin\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443162 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443251 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cni-binary-copy\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443331 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-multus-certs\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443413 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-os-release\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443494 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-bin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443578 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443680 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69fd7067-ac62-426d-b3a3-dca2750d14db-proxy-tls\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443768 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69fd7067-ac62-426d-b3a3-dca2750d14db-mcd-auth-proxy-config\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443848 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wrpp\" (UniqueName: \"kubernetes.io/projected/e83f8b9c-424d-4983-b5d3-83b52939c5eb-kube-api-access-6wrpp\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443923 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.443993 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444081 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444165 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.448648 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-multus\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444454 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-socket-dir-parent\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444480 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-system-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444496 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-netns\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.445158 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.445586 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.445786 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.445862 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6900ee15-3789-4e35-8d10-573acac2f5bd-hosts-file\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446497 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-daemon-config\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446516 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-conf-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446930 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-system-cni-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446947 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-k8s-cni-cncf-io\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446962 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.446977 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447159 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-multus-cni-dir\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.441943 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447178 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442028 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-etc-kubernetes\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447195 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.442059 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447211 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447239 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-os-release\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447256 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-kubelet\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447253 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-hostroot\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447649 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.444289 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.448982 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447707 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-os-release\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447718 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447735 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cnibin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447694 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-var-lib-cni-bin\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449058 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.448494 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cnibin\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.448643 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88k4h\" (UniqueName: \"kubernetes.io/projected/6900ee15-3789-4e35-8d10-573acac2f5bd-kube-api-access-88k4h\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449188 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449204 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449217 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449230 4688 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449241 4688 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449251 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449261 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449272 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449298 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449309 4688 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449320 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449331 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449342 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449354 4688 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449364 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449375 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449387 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449398 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449410 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449420 4688 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449432 4688 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449443 4688 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449454 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449465 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449477 4688 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449488 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449498 4688 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449509 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449519 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449547 4688 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449558 4688 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449569 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449582 4688 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449592 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449603 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449614 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449618 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-cni-binary-copy\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447673 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.447749 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450185 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/69fd7067-ac62-426d-b3a3-dca2750d14db-mcd-auth-proxy-config\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.448338 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.449625 4688 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450232 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450242 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450250 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450261 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450269 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450277 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450286 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450295 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450303 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450312 4688 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450320 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450328 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450337 4688 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450346 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450356 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450364 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450372 4688 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450383 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450392 4688 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450404 4688 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450414 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450425 4688 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450435 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450446 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450457 4688 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450467 4688 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450478 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450489 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450500 4688 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450511 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450542 4688 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450552 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450563 4688 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450575 4688 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450589 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450599 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450609 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450620 4688 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450631 4688 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450639 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450650 4688 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450659 4688 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450667 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450675 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450683 4688 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450693 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450701 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450709 4688 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450718 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450725 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450734 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450743 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450751 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450759 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450767 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450777 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450785 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450793 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450793 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/69fd7067-ac62-426d-b3a3-dca2750d14db-proxy-tls\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450802 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450840 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450852 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450863 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450873 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450881 4688 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450890 4688 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450898 4688 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450907 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450917 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450925 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450934 4688 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450942 4688 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450952 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450960 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450968 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450977 4688 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450985 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.450995 4688 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451005 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451013 4688 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451022 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451031 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451039 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451048 4688 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451056 4688 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451050 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451065 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451086 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451092 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451097 4688 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451109 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451120 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451130 4688 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451137 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-host-run-multus-certs\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451141 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451160 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.451403 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.452247 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e83f8b9c-424d-4983-b5d3-83b52939c5eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.452483 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.453139 4688 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.453175 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.453180 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.453691 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e83f8b9c-424d-4983-b5d3-83b52939c5eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.454000 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.454652 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.456983 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.457414 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.458034 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462223 4688 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462267 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462280 4688 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462290 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462299 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462308 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462319 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462329 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462339 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462347 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462355 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462364 4688 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462373 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462384 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462392 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462401 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462410 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462728 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462833 4688 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462845 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462853 4688 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462857 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.463307 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.462862 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.463769 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.463853 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.463931 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464009 4688 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464089 4688 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464165 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464243 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464300 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464298 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464436 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464540 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464623 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464699 4688 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464753 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464865 4688 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.464951 4688 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465025 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465096 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465154 4688 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465233 4688 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465309 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465372 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7cn\" (UniqueName: \"kubernetes.io/projected/69fd7067-ac62-426d-b3a3-dca2750d14db-kube-api-access-lq7cn\") pod \"machine-config-daemon-6ndlj\" (UID: \"69fd7067-ac62-426d-b3a3-dca2750d14db\") " pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465402 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465388 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465615 4688 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465630 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465645 4688 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465658 4688 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465670 4688 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465683 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465696 4688 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465708 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465720 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465731 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465743 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465755 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465767 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465778 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465790 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465804 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465819 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.465832 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.466417 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.466597 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.467294 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.467941 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.469281 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wrpp\" (UniqueName: \"kubernetes.io/projected/e83f8b9c-424d-4983-b5d3-83b52939c5eb-kube-api-access-6wrpp\") pod \"multus-additional-cni-plugins-nrzjb\" (UID: \"e83f8b9c-424d-4983-b5d3-83b52939c5eb\") " pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.475756 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.476258 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.478237 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.478708 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.479157 4688 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.483641 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.485789 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.486391 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.490081 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8bsc\" (UniqueName: \"kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc\") pod \"ovnkube-node-xgf9s\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.490604 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88k4h\" (UniqueName: \"kubernetes.io/projected/6900ee15-3789-4e35-8d10-573acac2f5bd-kube-api-access-88k4h\") pod \"node-resolver-xq8gs\" (UID: \"6900ee15-3789-4e35-8d10-573acac2f5bd\") " pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.490875 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2p6x\" (UniqueName: \"kubernetes.io/projected/ac2218c2-479a-4efd-9c1b-ed0d3968b06b-kube-api-access-j2p6x\") pod \"multus-jcfs4\" (UID: \"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\") " pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.490966 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.492953 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.493761 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.496813 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.499980 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.500721 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.501794 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.502321 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.507872 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.508450 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.509385 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.509864 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.510732 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.511347 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.518474 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.519264 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.520767 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.520872 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.521198 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.521581 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.521651 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.522852 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.523745 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.524227 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.529910 4688 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721" exitCode=255 Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.530114 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721"} Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.530233 4688 scope.go:117] "RemoveContainer" containerID="45e421840a3ff25e9ed20ff4f9b61565aeb91c478abf9613398df44c267c3c57" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.539752 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.561964 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.567268 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.567294 4688 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.572789 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.584879 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.593656 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.607298 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.608342 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.616561 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.620371 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.621752 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.622641 4688 scope.go:117] "RemoveContainer" containerID="7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.622842 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.623647 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.631994 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jcfs4" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.639688 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.655907 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.664450 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.684273 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.697799 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.698043 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.709485 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.721078 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.723917 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xq8gs" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.727174 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:47:13 crc kubenswrapper[4688]: W1001 15:47:13.754980 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode83f8b9c_424d_4983_b5d3_83b52939c5eb.slice/crio-9f101a28a39ca069510519e12ad3c7bd105b7cde9a0019d769de86dc69c267a5 WatchSource:0}: Error finding container 9f101a28a39ca069510519e12ad3c7bd105b7cde9a0019d769de86dc69c267a5: Status 404 returned error can't find the container with id 9f101a28a39ca069510519e12ad3c7bd105b7cde9a0019d769de86dc69c267a5 Oct 01 15:47:13 crc kubenswrapper[4688]: W1001 15:47:13.759722 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6900ee15_3789_4e35_8d10_573acac2f5bd.slice/crio-6eb98ba245a879b08b6ee0e0b8dc7877d21ff46c4234c0f0716be599a7c1412f WatchSource:0}: Error finding container 6eb98ba245a879b08b6ee0e0b8dc7877d21ff46c4234c0f0716be599a7c1412f: Status 404 returned error can't find the container with id 6eb98ba245a879b08b6ee0e0b8dc7877d21ff46c4234c0f0716be599a7c1412f Oct 01 15:47:13 crc kubenswrapper[4688]: W1001 15:47:13.762768 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69fd7067_ac62_426d_b3a3_dca2750d14db.slice/crio-43871a3b34a550f00fa00e86c8d007ceb353de164c3ef772a1f3ebd42d940a38 WatchSource:0}: Error finding container 43871a3b34a550f00fa00e86c8d007ceb353de164c3ef772a1f3ebd42d940a38: Status 404 returned error can't find the container with id 43871a3b34a550f00fa00e86c8d007ceb353de164c3ef772a1f3ebd42d940a38 Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.874842 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.874976 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:14.874927811 +0000 UTC m=+24.225567773 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.875233 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.875261 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.875343 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.875375 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:14.875367694 +0000 UTC m=+24.226007656 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.875402 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.875421 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:14.875414745 +0000 UTC m=+24.226054707 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.976212 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:13 crc kubenswrapper[4688]: I1001 15:47:13.976275 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976419 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976439 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976452 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976503 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:14.976486565 +0000 UTC m=+24.327126527 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976870 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976894 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976904 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:13 crc kubenswrapper[4688]: E1001 15:47:13.976936 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:14.976925617 +0000 UTC m=+24.327565589 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.534951 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.538137 4688 scope.go:117] "RemoveContainer" containerID="7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721" Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.538383 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.539873 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xq8gs" event={"ID":"6900ee15-3789-4e35-8d10-573acac2f5bd","Type":"ContainerStarted","Data":"8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.539919 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xq8gs" event={"ID":"6900ee15-3789-4e35-8d10-573acac2f5bd","Type":"ContainerStarted","Data":"6eb98ba245a879b08b6ee0e0b8dc7877d21ff46c4234c0f0716be599a7c1412f"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.541388 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerStarted","Data":"329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.541418 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerStarted","Data":"704c906c8b72dbd7b30853ea63cc882e5d4c0878f902830b904c444a8acbbe9e"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.544217 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.544249 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.544261 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3999a621cf7f48700786838c9199bd5fc137079c506db268dbc76ebac064c9a2"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.546589 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.546633 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f9b06547297314706df01c28b11b0ea4af200c917050923f17e23b58d2971400"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.548966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.549001 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.549018 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"43871a3b34a550f00fa00e86c8d007ceb353de164c3ef772a1f3ebd42d940a38"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.550805 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" exitCode=0 Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.550871 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.550901 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"cc3060133fd8fb15250beefd6ec6f1675e35047ef3ea93e73136f976134b3735"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.552786 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"33d108b25f36cdf3fd9167307901b34b5dddd61030ae95b7d733775bb8cb000a"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.554792 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83" exitCode=0 Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.554832 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.554878 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerStarted","Data":"9f101a28a39ca069510519e12ad3c7bd105b7cde9a0019d769de86dc69c267a5"} Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.555411 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.577065 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.592177 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.601049 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.620842 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.634715 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.669952 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.695541 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.713524 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.722502 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.722568 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.733983 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.749208 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.762865 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.775882 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.796837 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.811638 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.826075 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.836873 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.893380 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.893461 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.893490 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.893600 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.893645 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:16.893632966 +0000 UTC m=+26.244272928 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.893659 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.893720 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:16.893680188 +0000 UTC m=+26.244320150 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.893750 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:16.893742819 +0000 UTC m=+26.244382781 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.894512 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.920468 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.930495 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.948312 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.962103 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.979693 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.993944 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.993981 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994070 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994087 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994097 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994131 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:16.9941196 +0000 UTC m=+26.344759562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994136 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994158 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994170 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:14 crc kubenswrapper[4688]: E1001 15:47:14.994217 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:16.994202212 +0000 UTC m=+26.344842174 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:14 crc kubenswrapper[4688]: I1001 15:47:14.996947 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:14Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.010334 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.022974 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.380825 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:15 crc kubenswrapper[4688]: E1001 15:47:15.381231 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.380867 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.380865 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:15 crc kubenswrapper[4688]: E1001 15:47:15.381298 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:15 crc kubenswrapper[4688]: E1001 15:47:15.381487 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.386881 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.387833 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.559627 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b" exitCode=0 Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.559701 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565516 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565822 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565838 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565852 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565863 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.565874 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.575375 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.594421 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.608296 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.620116 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.639156 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.654733 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.665036 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.676646 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.690205 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.701202 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.714764 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.729674 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:15 crc kubenswrapper[4688]: I1001 15:47:15.751143 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:15Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.570674 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2" exitCode=0 Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.570773 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2"} Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.572583 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003"} Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.592988 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.611686 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.638757 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.656119 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.675114 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.696296 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.711972 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.728363 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.746721 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.763498 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.776349 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.789233 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.800668 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.817790 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.833098 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.852367 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.870803 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.886665 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.899419 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.911468 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.911910 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:16 crc kubenswrapper[4688]: E1001 15:47:16.912050 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:20.912033037 +0000 UTC m=+30.262672999 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.912111 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.912153 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:16 crc kubenswrapper[4688]: E1001 15:47:16.912246 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:16 crc kubenswrapper[4688]: E1001 15:47:16.912305 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:20.912291225 +0000 UTC m=+30.262931197 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:16 crc kubenswrapper[4688]: E1001 15:47:16.912252 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:16 crc kubenswrapper[4688]: E1001 15:47:16.912378 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:20.912367607 +0000 UTC m=+30.263007569 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.923319 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.939871 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.954541 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.967639 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.979422 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:16 crc kubenswrapper[4688]: I1001 15:47:16.990664 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:16Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.013096 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.013147 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013276 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013277 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013291 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013303 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013332 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013337 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013377 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:21.013364015 +0000 UTC m=+30.364003977 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.013391 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:21.013386305 +0000 UTC m=+30.364026267 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.380679 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.380808 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.380833 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.380869 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.380932 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:17 crc kubenswrapper[4688]: E1001 15:47:17.380986 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.583074 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf" exitCode=0 Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.583158 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf"} Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.601297 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.603015 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.622407 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.635835 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.653267 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.668077 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.683157 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.699833 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.717335 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.735468 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.749768 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.761542 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.782171 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:17 crc kubenswrapper[4688]: I1001 15:47:17.797147 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:17Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.109232 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ccw98"] Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.109570 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.112828 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.112987 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.113203 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.113378 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.135194 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.151131 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.164945 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.186829 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.227341 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pt8\" (UniqueName: \"kubernetes.io/projected/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-kube-api-access-l9pt8\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.227432 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-serviceca\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.227476 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-host\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.228484 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.249484 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.270054 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.284949 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.303421 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.319087 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.330956 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-serviceca\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.331026 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-host\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.331076 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pt8\" (UniqueName: \"kubernetes.io/projected/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-kube-api-access-l9pt8\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.331501 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-host\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.332228 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-serviceca\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.345375 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.361211 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.367160 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pt8\" (UniqueName: \"kubernetes.io/projected/94ed6916-e5b4-40b9-87ed-fee49ecb1c28-kube-api-access-l9pt8\") pod \"node-ca-ccw98\" (UID: \"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\") " pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.384995 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.396978 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.430108 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ccw98" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.613036 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda" exitCode=0 Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.613201 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.615681 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ccw98" event={"ID":"94ed6916-e5b4-40b9-87ed-fee49ecb1c28","Type":"ContainerStarted","Data":"2a9f6585bcb436771c74202b0a3bd33cc9b41f4e266d1877a5f78893ec707293"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.640493 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.656613 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.681063 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.695729 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.712577 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.725781 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.727740 4688 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.730556 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.730622 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.730638 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.730832 4688 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.737929 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.738913 4688 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.739152 4688 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.751743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.751781 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.751793 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.751811 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.751824 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.755770 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.765850 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.768579 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.769124 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.769162 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.769174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.769192 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.769203 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.782023 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.784974 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.785498 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.785518 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.785546 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.785559 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.785567 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.800942 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.807229 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.811353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.811388 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.811401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.811418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.811432 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.812762 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.823779 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.826615 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.827085 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.827110 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.827122 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.827138 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.827149 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.837265 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.839633 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:18Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:18 crc kubenswrapper[4688]: E1001 15:47:18.839737 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.841015 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.841068 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.841080 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.841095 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.841107 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.944039 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.944079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.944092 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.944108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:18 crc kubenswrapper[4688]: I1001 15:47:18.944119 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:18Z","lastTransitionTime":"2025-10-01T15:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.046724 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.047128 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.047314 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.047460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.047618 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.150195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.150284 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.150307 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.150346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.150369 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.253291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.253338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.253352 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.253372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.253388 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.357407 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.357471 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.357488 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.357512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.357558 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.381708 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.382505 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.381709 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:19 crc kubenswrapper[4688]: E1001 15:47:19.383128 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:19 crc kubenswrapper[4688]: E1001 15:47:19.383234 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:19 crc kubenswrapper[4688]: E1001 15:47:19.383451 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.460722 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.460777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.460802 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.460825 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.460841 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.570134 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.570220 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.570242 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.570272 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.570292 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.621032 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ccw98" event={"ID":"94ed6916-e5b4-40b9-87ed-fee49ecb1c28","Type":"ContainerStarted","Data":"ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.633373 4688 generic.go:334] "Generic (PLEG): container finished" podID="e83f8b9c-424d-4983-b5d3-83b52939c5eb" containerID="6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a" exitCode=0 Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.633440 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerDied","Data":"6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.638702 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.651727 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.667311 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.674455 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.674505 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.674547 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.674571 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.674587 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.698082 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.714748 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.736500 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.746263 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.761329 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.772129 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.776629 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.776675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.776691 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.776712 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.776728 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.785161 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.796213 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.806372 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.822272 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.839055 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.849600 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.864220 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.878289 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.881412 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.881463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.881474 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.881493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.881505 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.893872 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.907254 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.924874 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.936991 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.950863 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.965400 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.979414 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.987875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.987921 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.987932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.987949 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.987992 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:19Z","lastTransitionTime":"2025-10-01T15:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:19 crc kubenswrapper[4688]: I1001 15:47:19.992968 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:19Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.005027 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.016963 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.027069 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.090296 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.090322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.090329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.090341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.090350 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.199274 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.199302 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.199310 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.199322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.199331 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.301894 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.301929 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.301939 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.301951 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.301960 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.404460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.404499 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.404508 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.404538 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.404548 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.507687 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.507735 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.507751 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.507776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.507796 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.610190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.610245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.610259 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.610277 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.610288 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.645907 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.646206 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.646256 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.646285 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.656630 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" event={"ID":"e83f8b9c-424d-4983-b5d3-83b52939c5eb","Type":"ContainerStarted","Data":"6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.669949 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.688690 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.688816 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.694905 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.713208 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.713284 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.713304 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.713328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.713346 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.716698 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.741567 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.756427 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.773995 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.790140 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.805631 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.816951 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.817010 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.817027 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.817050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.817068 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.823373 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.841348 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.858087 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.881272 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.902158 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.915366 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.919130 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.919188 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.919207 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.919230 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.919244 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:20Z","lastTransitionTime":"2025-10-01T15:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.932576 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.943564 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.964239 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.966355 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.966471 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:20 crc kubenswrapper[4688]: E1001 15:47:20.966514 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:28.966488712 +0000 UTC m=+38.317128674 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.966588 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:20 crc kubenswrapper[4688]: E1001 15:47:20.966615 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:20 crc kubenswrapper[4688]: E1001 15:47:20.966665 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:28.966650186 +0000 UTC m=+38.317290248 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:20 crc kubenswrapper[4688]: E1001 15:47:20.966672 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:20 crc kubenswrapper[4688]: E1001 15:47:20.966714 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:28.966707008 +0000 UTC m=+38.317346970 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.979285 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:20 crc kubenswrapper[4688]: I1001 15:47:20.996290 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.006878 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.019579 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.021370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.021399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.021410 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.021425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.021436 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.034451 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.052174 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.066870 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.067134 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.067208 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067300 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067324 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067337 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067368 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067387 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:29.067370936 +0000 UTC m=+38.418010898 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067392 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067409 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.067483 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:29.067454789 +0000 UTC m=+38.418094781 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.080121 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.096463 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.108235 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.120014 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.123997 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.124050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.124067 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.124091 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.124109 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.226852 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.226920 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.226942 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.226969 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.226992 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.329977 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.330024 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.330035 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.330060 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.330076 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.380425 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.380556 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.380648 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.380876 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.380981 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.381138 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.397153 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.411735 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.425786 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.434671 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.434730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.434743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.434759 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.434769 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.449803 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.466925 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.484592 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.501683 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.514197 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.528599 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537404 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537442 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537451 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537465 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537475 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.537808 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.551324 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.569605 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.584981 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.597881 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.639761 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.639804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.639818 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.639838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.639853 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.743339 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.744190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.744284 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.744370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.744456 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.847696 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.847797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.847867 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.847892 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.847947 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.951622 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.951835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.951923 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.951990 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.952052 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:21Z","lastTransitionTime":"2025-10-01T15:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.970664 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:21 crc kubenswrapper[4688]: I1001 15:47:21.971897 4688 scope.go:117] "RemoveContainer" containerID="7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721" Oct 01 15:47:21 crc kubenswrapper[4688]: E1001 15:47:21.972192 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.054705 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.055115 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.055518 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.055977 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.056289 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.161222 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.161595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.162019 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.162293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.162480 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.265659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.265725 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.265750 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.265778 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.265800 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.399436 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.399500 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.399557 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.399582 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.399609 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.501481 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.501509 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.501531 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.501545 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.501553 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.603809 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.603853 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.603865 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.603883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.603894 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.705351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.705384 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.705392 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.705405 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.705415 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.807997 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.808036 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.808045 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.808059 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.808068 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.910294 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.910342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.910353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.910371 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:22 crc kubenswrapper[4688]: I1001 15:47:22.910383 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:22Z","lastTransitionTime":"2025-10-01T15:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.012176 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.012219 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.012267 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.012289 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.012305 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.115308 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.115356 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.115373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.115396 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.115413 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.218740 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.218808 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.218826 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.218852 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.218870 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.322681 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.322772 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.322791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.322815 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.322832 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.380683 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.380752 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:23 crc kubenswrapper[4688]: E1001 15:47:23.380913 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:23 crc kubenswrapper[4688]: E1001 15:47:23.381083 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.380710 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:23 crc kubenswrapper[4688]: E1001 15:47:23.381267 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.425589 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.425644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.425665 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.425693 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.425714 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.529160 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.529638 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.529679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.529712 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.529734 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.633549 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.633932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.634123 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.634313 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.634519 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.668357 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/0.log" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.674064 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa" exitCode=1 Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.674132 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.675461 4688 scope.go:117] "RemoveContainer" containerID="baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.698707 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.717063 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.732836 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.737677 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.737726 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.737745 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.737765 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.737781 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.749051 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.768557 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.787301 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.804187 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.819573 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.840162 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.840238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.840253 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.840270 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.840282 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.846616 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.863987 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.893930 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.908630 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.935363 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.950613 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.950657 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.950668 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.950686 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.950699 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:23Z","lastTransitionTime":"2025-10-01T15:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:23 crc kubenswrapper[4688]: I1001 15:47:23.954934 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:23Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.053108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.053145 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.053156 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.053171 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.053183 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.155319 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.155346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.155354 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.155366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.155375 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.257431 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.257457 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.257465 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.257478 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.257488 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.360028 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.360081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.360101 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.360124 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.360139 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.463254 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.463320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.463336 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.463361 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.463378 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.568278 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.568330 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.568352 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.568380 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.568403 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.672030 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.672082 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.672095 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.672113 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.672126 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.680627 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/1.log" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.681616 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/0.log" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.685597 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1" exitCode=1 Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.685652 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.685709 4688 scope.go:117] "RemoveContainer" containerID="baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.686898 4688 scope.go:117] "RemoveContainer" containerID="3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1" Oct 01 15:47:24 crc kubenswrapper[4688]: E1001 15:47:24.687147 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.711132 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.723908 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.740561 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.765003 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.775401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.775432 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.775445 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.775465 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.775481 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.780324 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.805440 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.822313 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.853512 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.871693 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.878051 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.878098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.878118 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.878141 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.878159 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.889574 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.905256 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.922627 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.940890 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.959549 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:24Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.981107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.981164 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.981180 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.981203 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:24 crc kubenswrapper[4688]: I1001 15:47:24.981219 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:24Z","lastTransitionTime":"2025-10-01T15:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.084445 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.084497 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.084516 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.084570 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.084587 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.187270 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.187322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.187339 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.187390 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.187407 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.290626 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.291054 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.291228 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.291373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.291604 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.379960 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.379992 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.379993 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:25 crc kubenswrapper[4688]: E1001 15:47:25.380158 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:25 crc kubenswrapper[4688]: E1001 15:47:25.380243 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:25 crc kubenswrapper[4688]: E1001 15:47:25.380313 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.396058 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.396105 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.396117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.396135 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.396149 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.499329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.499392 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.499411 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.499439 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.499457 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.602151 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.602611 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.602750 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.602932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.603107 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.692626 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/1.log" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.705965 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.706184 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.706324 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.706470 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.706647 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.810307 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.810644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.810792 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.810950 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.811091 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.821476 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6"] Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.822408 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.825671 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.826076 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.849958 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.867704 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.883379 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.900241 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.913746 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.913819 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.913845 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.913875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.913899 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:25Z","lastTransitionTime":"2025-10-01T15:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.921786 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.924458 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55lj6\" (UniqueName: \"kubernetes.io/projected/1c2b8d9c-d29c-497d-bac1-3eba88343919-kube-api-access-55lj6\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.924611 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.924660 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-env-overrides\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.924698 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.943771 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.964360 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.983263 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:25 crc kubenswrapper[4688]: I1001 15:47:25.998210 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:25Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.017056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.017128 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.017148 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.017172 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.017191 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.026380 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55lj6\" (UniqueName: \"kubernetes.io/projected/1c2b8d9c-d29c-497d-bac1-3eba88343919-kube-api-access-55lj6\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.026481 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.026561 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-env-overrides\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.026608 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.028043 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.029368 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c2b8d9c-d29c-497d-bac1-3eba88343919-env-overrides\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.033254 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c2b8d9c-d29c-497d-bac1-3eba88343919-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.031966 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.046686 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55lj6\" (UniqueName: \"kubernetes.io/projected/1c2b8d9c-d29c-497d-bac1-3eba88343919-kube-api-access-55lj6\") pod \"ovnkube-control-plane-749d76644c-trwg6\" (UID: \"1c2b8d9c-d29c-497d-bac1-3eba88343919\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.059857 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.084226 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.099891 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.117919 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.120343 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.120387 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.120401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.120418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.120429 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.136647 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.143719 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.223461 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.223512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.223560 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.223584 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.223602 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.326605 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.326638 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.326646 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.326659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.326668 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.430319 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.430343 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.430351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.430364 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.430373 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.538027 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.538058 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.538066 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.538079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.538088 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.642023 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.642081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.642099 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.642125 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.642142 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.701213 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" event={"ID":"1c2b8d9c-d29c-497d-bac1-3eba88343919","Type":"ContainerStarted","Data":"c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.701284 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" event={"ID":"1c2b8d9c-d29c-497d-bac1-3eba88343919","Type":"ContainerStarted","Data":"caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.701305 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" event={"ID":"1c2b8d9c-d29c-497d-bac1-3eba88343919","Type":"ContainerStarted","Data":"35e33d9d124c3617fee7daacfefacebaf2a10a05b1fd1fd4ec8637d86ca83efe"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.720851 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.735386 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.745076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.745131 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.745146 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.745169 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.745184 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.751784 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.772120 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.789939 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.806257 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.819249 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.833091 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.846256 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.848878 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.848932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.848952 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.848974 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.848993 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.859326 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.902771 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.918229 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.942304 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.952392 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.952661 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.952723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.952808 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.952866 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:26Z","lastTransitionTime":"2025-10-01T15:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.957994 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:26 crc kubenswrapper[4688]: I1001 15:47:26.973197 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:26Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.026350 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-5sm44"] Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.027313 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.027571 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.042481 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.055644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.055707 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.055724 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.055743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.055755 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.056064 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.070021 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.080981 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.081217 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gwf\" (UniqueName: \"kubernetes.io/projected/c987bdd3-e8bc-473e-99b2-61a143a95543-kube-api-access-78gwf\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.084378 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.094814 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.107129 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.131443 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.145321 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.158263 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.158381 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.158475 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.158587 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.158665 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.166964 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.182316 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gwf\" (UniqueName: \"kubernetes.io/projected/c987bdd3-e8bc-473e-99b2-61a143a95543-kube-api-access-78gwf\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.182443 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.182614 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.182693 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:27.68267237 +0000 UTC m=+37.033312342 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.183139 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.210912 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.211048 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gwf\" (UniqueName: \"kubernetes.io/projected/c987bdd3-e8bc-473e-99b2-61a143a95543-kube-api-access-78gwf\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.225662 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.238014 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.249469 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.261169 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.261353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.261542 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.261688 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.261818 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.266044 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.278476 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:27Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.365274 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.365307 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.365315 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.365329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.365338 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.380016 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.380092 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.380103 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.380194 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.380278 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.380325 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.468240 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.468299 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.468317 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.468341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.468358 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.571061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.571129 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.571148 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.571173 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.571191 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.674453 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.674585 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.674609 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.674637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.674662 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.686898 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.687115 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:27 crc kubenswrapper[4688]: E1001 15:47:27.687208 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:28.687184836 +0000 UTC m=+38.037824828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.777900 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.777973 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.777990 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.778009 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.778023 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.881117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.881169 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.881181 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.881197 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.881211 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.983317 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.983348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.983358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.983371 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:27 crc kubenswrapper[4688]: I1001 15:47:27.983381 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:27Z","lastTransitionTime":"2025-10-01T15:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.086809 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.086887 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.086914 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.086944 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.087343 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.189684 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.189740 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.189749 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.189764 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.189773 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.294271 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.294590 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.294606 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.294629 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.294640 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.380183 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.380361 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.397302 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.397393 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.397412 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.397437 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.397453 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.499304 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.499352 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.499366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.499386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.499402 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.602619 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.602681 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.602700 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.602723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.602741 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.694289 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.694503 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.694683 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:30.694646182 +0000 UTC m=+40.045286184 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.706918 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.706954 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.706964 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.706997 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.707009 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.809188 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.809227 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.809235 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.809248 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.809258 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.912943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.913004 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.913020 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.913043 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.913061 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:28Z","lastTransitionTime":"2025-10-01T15:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.996308 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.996444 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:47:44.99641353 +0000 UTC m=+54.347053522 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.996558 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:28 crc kubenswrapper[4688]: I1001 15:47:28.996624 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.996751 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.996789 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.996854 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:44.996831422 +0000 UTC m=+54.347471424 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:28 crc kubenswrapper[4688]: E1001 15:47:28.996888 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:44.996871933 +0000 UTC m=+54.347511945 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.016453 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.016569 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.016589 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.016618 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.016638 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.083190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.083287 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.083306 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.083368 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.083388 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.097011 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.097083 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097299 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097327 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097345 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097399 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097462 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097490 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097416 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:45.097396217 +0000 UTC m=+54.448036219 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.097665 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:45.097629094 +0000 UTC m=+54.448269096 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.103490 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:29Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.108259 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.108341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.108359 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.108379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.108396 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.127907 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:29Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.132435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.132493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.132519 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.132576 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.132598 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.161786 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:29Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.166294 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.166371 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.166395 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.166425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.166450 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.189100 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:29Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.193212 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.193242 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.193252 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.193268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.193279 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.214959 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:29Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.215137 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.217205 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.217234 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.217244 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.217260 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.217270 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.319384 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.319439 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.319463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.319493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.319516 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.380487 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.380694 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.381615 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.381711 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.381823 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:29 crc kubenswrapper[4688]: E1001 15:47:29.381984 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.422267 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.422338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.422350 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.422367 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.422381 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.525717 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.525780 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.525805 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.525835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.525857 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.629194 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.629268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.629293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.629321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.629343 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.731868 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.731919 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.731936 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.731957 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.731973 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.834349 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.834468 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.834493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.834553 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.834580 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.937320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.937379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.937396 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.937418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:29 crc kubenswrapper[4688]: I1001 15:47:29.937439 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:29Z","lastTransitionTime":"2025-10-01T15:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.040727 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.040797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.040813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.040842 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.040860 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.143930 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.143998 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.144016 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.144041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.144059 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.247232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.247303 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.247328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.247357 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.247380 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.350942 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.350999 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.351015 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.351037 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.351053 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.380821 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:30 crc kubenswrapper[4688]: E1001 15:47:30.381026 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.453643 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.453679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.453688 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.453701 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.453710 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.557720 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.557777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.557802 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.557831 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.557849 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.660831 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.660888 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.660905 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.660929 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.660946 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.715085 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:30 crc kubenswrapper[4688]: E1001 15:47:30.715259 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:30 crc kubenswrapper[4688]: E1001 15:47:30.715363 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:34.715335987 +0000 UTC m=+44.065975989 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.764467 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.764856 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.765090 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.765291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.765495 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.868343 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.868375 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.868402 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.868417 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.868425 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.971463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.971497 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.971538 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.971554 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:30 crc kubenswrapper[4688]: I1001 15:47:30.971563 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:30Z","lastTransitionTime":"2025-10-01T15:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.074282 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.074631 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.075238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.075293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.075332 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.192898 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.192959 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.192976 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.192998 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.193014 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.295901 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.295965 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.296028 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.296054 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.296074 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.380694 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.380766 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.381283 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:31 crc kubenswrapper[4688]: E1001 15:47:31.381547 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:31 crc kubenswrapper[4688]: E1001 15:47:31.381656 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:31 crc kubenswrapper[4688]: E1001 15:47:31.382428 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.399571 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.399627 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.399644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.399667 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.399687 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.405773 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.422226 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.440680 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.472328 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baf98441ea342540966861407bab5d65f56bac0e72c9b74b115f4ed13ee7c8fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:23Z\\\",\\\"message\\\":\\\"1.Namespace event handler 5\\\\nI1001 15:47:22.979581 5938 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 15:47:22.979587 5938 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979644 5938 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.979769 5938 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 15:47:22.980275 5938 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.980960 5938 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 15:47:22.981610 5938 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982404 5938 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 15:47:22.982662 5938 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.492959 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.502353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.502558 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.502720 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.502845 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.502965 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.512279 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.523993 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.539974 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.552615 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.561945 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.576098 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.589441 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.599478 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.606115 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.606168 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.606180 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.606197 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.606208 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.612002 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.625224 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.636463 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:31Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.709252 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.709300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.709317 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.709338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.709356 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.821076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.821169 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.821194 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.821227 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.821250 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.925360 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.925426 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.925443 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.925474 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:31 crc kubenswrapper[4688]: I1001 15:47:31.925498 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:31Z","lastTransitionTime":"2025-10-01T15:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.028243 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.028360 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.028381 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.028403 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.028420 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.130732 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.130772 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.130787 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.130807 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.130824 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.234053 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.234107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.234128 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.234155 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.234176 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.337719 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.337782 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.337804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.337833 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.337858 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.380140 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:32 crc kubenswrapper[4688]: E1001 15:47:32.380374 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.440647 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.440700 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.440723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.440753 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.440777 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.543416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.543484 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.543501 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.543581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.543606 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.646297 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.646330 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.646338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.646354 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.646368 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.748779 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.748856 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.748878 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.748906 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.748930 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.852016 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.852088 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.852107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.852146 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.852166 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.954818 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.954896 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.954937 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.954967 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:32 crc kubenswrapper[4688]: I1001 15:47:32.954989 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:32Z","lastTransitionTime":"2025-10-01T15:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.057788 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.057862 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.057880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.057910 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.057929 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.161723 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.161848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.161865 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.161890 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.161907 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.265241 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.265437 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.265462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.265491 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.265514 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.369005 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.369087 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.369113 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.369145 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.369165 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.380756 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.380790 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:33 crc kubenswrapper[4688]: E1001 15:47:33.380897 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.380963 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:33 crc kubenswrapper[4688]: E1001 15:47:33.381023 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:33 crc kubenswrapper[4688]: E1001 15:47:33.381207 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.471737 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.471798 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.471819 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.471845 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.471868 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.575034 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.575108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.575133 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.575160 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.575178 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.678321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.678399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.678416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.678446 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.678463 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.781706 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.781800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.781876 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.781901 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.781922 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.884320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.884406 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.884455 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.884477 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.884496 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.987828 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.987919 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.987943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.987971 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:33 crc kubenswrapper[4688]: I1001 15:47:33.987999 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:33Z","lastTransitionTime":"2025-10-01T15:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.091135 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.091248 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.091276 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.091303 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.091323 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.194197 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.194253 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.194270 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.194291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.194309 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.297929 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.297996 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.298013 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.298042 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.298060 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.380448 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:34 crc kubenswrapper[4688]: E1001 15:47:34.380765 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.402689 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.402756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.402874 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.402909 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.402932 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.506437 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.506503 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.506563 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.506601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.506624 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.608967 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.609136 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.609181 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.609213 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.609236 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.711806 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.712108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.712126 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.712149 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.712165 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.762880 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:34 crc kubenswrapper[4688]: E1001 15:47:34.763073 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:34 crc kubenswrapper[4688]: E1001 15:47:34.763198 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:42.763166493 +0000 UTC m=+52.113806495 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.815239 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.815329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.815346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.815371 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.815422 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.917476 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.917559 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.917578 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.917632 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:34 crc kubenswrapper[4688]: I1001 15:47:34.917651 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:34Z","lastTransitionTime":"2025-10-01T15:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.020443 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.020499 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.020508 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.020561 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.020573 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.123286 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.123348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.123366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.123389 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.123409 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.226279 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.226332 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.226348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.226366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.226379 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.329331 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.329425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.329439 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.329489 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.329505 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.380547 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.380701 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:35 crc kubenswrapper[4688]: E1001 15:47:35.380772 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.380605 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:35 crc kubenswrapper[4688]: E1001 15:47:35.380873 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:35 crc kubenswrapper[4688]: E1001 15:47:35.381018 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.432182 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.432233 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.432245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.432301 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.432315 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.535731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.535797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.535814 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.535838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.535855 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.639337 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.639386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.639398 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.639416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.639429 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.742682 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.742766 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.742785 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.742813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.742836 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.845692 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.845745 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.845764 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.845785 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.845837 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.948487 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.948577 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.948602 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.948630 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:35 crc kubenswrapper[4688]: I1001 15:47:35.948650 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:35Z","lastTransitionTime":"2025-10-01T15:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.051237 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.051300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.051324 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.051351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.051376 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.153841 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.153881 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.153892 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.153907 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.153919 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.257642 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.257693 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.257706 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.257725 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.257739 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.362679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.362777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.362804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.362835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.362857 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.379999 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:36 crc kubenswrapper[4688]: E1001 15:47:36.380988 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.381022 4688 scope.go:117] "RemoveContainer" containerID="7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.381461 4688 scope.go:117] "RemoveContainer" containerID="3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.406680 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.428695 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.443337 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.457865 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.466885 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.466932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.466943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.467596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.467617 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.467832 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.488641 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.503178 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.515957 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.528203 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.544258 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.555573 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.566037 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.574420 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.574473 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.574486 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.574516 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.574545 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.578615 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.598882 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.615688 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.629259 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.676609 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.676665 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.676682 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.676705 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.676724 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.738893 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/1.log" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.743252 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.743976 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.761302 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.779449 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.779506 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.779561 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.779593 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.779614 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.780768 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.797739 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.817083 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.846488 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.870567 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.881959 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.882025 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.882046 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.882071 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.882087 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.896715 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.916873 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.939581 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.957905 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.972929 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.983932 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.984763 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.984801 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.984810 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.984822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:36 crc kubenswrapper[4688]: I1001 15:47:36.984831 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:36Z","lastTransitionTime":"2025-10-01T15:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.001473 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:36Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.015689 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.028371 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.038377 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.087170 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.087207 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.087218 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.087234 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.087247 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.189752 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.189786 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.189795 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.189808 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.189819 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.292287 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.292321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.292344 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.292358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.292367 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.380459 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.380569 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.380606 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:37 crc kubenswrapper[4688]: E1001 15:47:37.380693 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:37 crc kubenswrapper[4688]: E1001 15:47:37.380815 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:37 crc kubenswrapper[4688]: E1001 15:47:37.381003 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.394179 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.394227 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.394238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.394253 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.394266 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.496050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.496081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.496097 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.496113 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.496122 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.598697 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.598756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.598773 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.598797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.598836 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.703297 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.703370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.703387 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.703414 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.703435 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.747445 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.749007 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.749406 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.751144 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/2.log" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.751959 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/1.log" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.755003 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" exitCode=1 Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.755049 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.755086 4688 scope.go:117] "RemoveContainer" containerID="3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.756055 4688 scope.go:117] "RemoveContainer" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" Oct 01 15:47:37 crc kubenswrapper[4688]: E1001 15:47:37.756272 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.768459 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.791292 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.806191 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.806222 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.806232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.806245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.806257 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.807678 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.828876 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.845463 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.866882 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.884566 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.904641 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.908546 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.908593 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.908610 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.908629 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.908646 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:37Z","lastTransitionTime":"2025-10-01T15:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.919426 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.934948 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.949727 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.968678 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:37 crc kubenswrapper[4688]: I1001 15:47:37.987190 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.005678 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.011592 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.011656 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.011675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.011700 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.011719 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.026975 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.040506 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.056050 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.104261 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d274e510b4c76ee6a361548f33046e4ba21daf5968c9bec6acfb69dad8957e1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:24Z\\\",\\\"message\\\":\\\"[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00705d47f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 15:47:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.113931 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.114267 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.114417 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.114595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.114748 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.124361 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.150176 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.162148 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.176840 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.187032 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.197999 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.209103 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.217269 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.217300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.217312 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.217328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.217339 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.221867 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.234249 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.250238 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.264935 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.278914 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.292057 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.304710 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.319494 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.319581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.319599 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.319621 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.319638 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.380584 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:38 crc kubenswrapper[4688]: E1001 15:47:38.380724 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.422634 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.422674 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.422682 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.422696 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.422705 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.525454 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.525569 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.525594 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.525624 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.525647 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.629481 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.629860 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.630068 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.630255 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.630401 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.733798 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.733869 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.733887 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.733914 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.733931 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.763363 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/2.log" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.770457 4688 scope.go:117] "RemoveContainer" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" Oct 01 15:47:38 crc kubenswrapper[4688]: E1001 15:47:38.770738 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.793616 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.815506 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.832132 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.836547 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.836575 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.836583 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.836600 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.836610 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.851384 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.871168 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.889775 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.904018 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.920552 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.940509 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.940561 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.940601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.940619 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.940631 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:38Z","lastTransitionTime":"2025-10-01T15:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.943592 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.955484 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.977490 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:38 crc kubenswrapper[4688]: I1001 15:47:38.995679 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:38Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.018539 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.034824 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.043605 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.043637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.043645 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.043660 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.043669 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.052742 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.064482 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.145850 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.145897 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.145914 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.145937 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.145954 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.235119 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.235172 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.235188 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.235210 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.235228 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.249809 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.254785 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.254838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.254856 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.254879 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.254896 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.273862 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.278961 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.279009 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.279027 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.279053 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.279071 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.298218 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.303254 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.303346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.303367 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.303391 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.303411 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.318549 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.323659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.323712 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.323730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.323756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.323775 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.345361 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:39Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.345631 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.350659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.350721 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.350742 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.350769 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.350795 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.380662 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.380770 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.380923 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.381121 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.384913 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:39 crc kubenswrapper[4688]: E1001 15:47:39.385137 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.454036 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.454093 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.454110 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.454135 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.454187 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.557313 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.557363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.557379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.557402 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.557418 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.661095 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.661146 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.661162 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.661184 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.661200 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.764487 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.764595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.764616 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.764644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.764663 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.867358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.867418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.867435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.867458 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.867476 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.976001 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.976058 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.976075 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.976101 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:39 crc kubenswrapper[4688]: I1001 15:47:39.976119 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:39Z","lastTransitionTime":"2025-10-01T15:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.079905 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.079963 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.079979 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.080005 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.080025 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.182900 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.182955 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.182968 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.182987 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.183002 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.285712 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.285776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.285798 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.285825 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.285843 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.380879 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:40 crc kubenswrapper[4688]: E1001 15:47:40.381096 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.390728 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.390812 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.390838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.390872 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.390896 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.494065 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.494149 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.494174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.494206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.494230 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.598082 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.598140 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.598158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.598181 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.598199 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.701138 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.701176 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.701187 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.701202 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.701213 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.803667 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.803733 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.803756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.803784 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.803806 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.907210 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.907280 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.907304 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.907333 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:40 crc kubenswrapper[4688]: I1001 15:47:40.907356 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:40Z","lastTransitionTime":"2025-10-01T15:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.010489 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.010599 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.010621 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.010644 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.010660 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.112963 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.113020 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.113043 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.113069 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.113094 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.215681 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.215724 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.215742 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.215763 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.215779 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.318717 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.318774 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.318791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.318813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.318834 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.380921 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.380984 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.380915 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:41 crc kubenswrapper[4688]: E1001 15:47:41.381087 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:41 crc kubenswrapper[4688]: E1001 15:47:41.381321 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:41 crc kubenswrapper[4688]: E1001 15:47:41.381455 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.400951 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.421829 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.422764 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.422873 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.422913 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.423143 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.423205 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.446473 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.469682 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.486307 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.509301 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.525702 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.527080 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.527119 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.527135 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.527156 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.527173 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.547952 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.565623 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.597481 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.615176 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.630107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.630183 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.630201 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.630220 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.630232 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.634431 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.648647 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.667506 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.687617 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.703599 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:41Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.733083 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.733150 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.733171 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.733201 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.733219 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.836124 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.836190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.836208 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.836234 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.836251 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.938817 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.938867 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.938883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.938907 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:41 crc kubenswrapper[4688]: I1001 15:47:41.938924 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:41Z","lastTransitionTime":"2025-10-01T15:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.041641 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.041699 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.041720 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.041744 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.041764 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.144883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.144931 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.144949 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.144972 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.144990 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.248184 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.248246 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.248264 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.248287 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.248305 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.351446 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.351513 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.351560 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.351585 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.351603 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.380006 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:42 crc kubenswrapper[4688]: E1001 15:47:42.380331 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.454275 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.454328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.454345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.454367 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.454387 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.559493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.559612 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.559637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.559665 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.559687 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.663288 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.663373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.663398 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.663431 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.663467 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.766409 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.766489 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.766515 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.766580 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.766605 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.855353 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:42 crc kubenswrapper[4688]: E1001 15:47:42.855682 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:42 crc kubenswrapper[4688]: E1001 15:47:42.855957 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:47:58.855925461 +0000 UTC m=+68.206565453 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.869562 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.869606 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.869626 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.869651 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.869668 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.972751 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.972797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.972813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.972835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:42 crc kubenswrapper[4688]: I1001 15:47:42.972851 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:42Z","lastTransitionTime":"2025-10-01T15:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.075655 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.075700 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.075716 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.075738 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.075754 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.178944 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.179011 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.179032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.179057 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.179075 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.281629 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.281702 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.281726 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.281756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.281778 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.380697 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.380759 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.380707 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:43 crc kubenswrapper[4688]: E1001 15:47:43.380872 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:43 crc kubenswrapper[4688]: E1001 15:47:43.381063 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:43 crc kubenswrapper[4688]: E1001 15:47:43.381204 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.384977 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.385011 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.385023 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.385040 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.385050 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.488667 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.488743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.488765 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.488794 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.488815 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.591672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.591739 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.591757 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.591782 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.591802 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.694674 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.694735 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.694939 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.694965 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.694984 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.797472 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.797571 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.797591 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.797615 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.797633 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.900706 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.900768 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.900785 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.900811 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:43 crc kubenswrapper[4688]: I1001 15:47:43.900831 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:43Z","lastTransitionTime":"2025-10-01T15:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.004059 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.004397 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.004568 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.004704 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.004825 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.107616 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.108071 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.108285 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.108500 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.108734 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.211640 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.211694 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.211711 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.211769 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.211793 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.314890 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.315224 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.315414 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.315605 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.315764 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.379981 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:44 crc kubenswrapper[4688]: E1001 15:47:44.380188 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.383090 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.397964 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.407172 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.418884 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.419322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.419347 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.419379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.419470 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.430159 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.451673 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.471456 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.488344 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.502682 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.521667 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.521716 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.521727 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.521745 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.521757 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.530963 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.547995 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.566441 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.582455 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.600269 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.615357 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.624210 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.624351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.624449 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.624562 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.624661 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.633911 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.647841 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.665614 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.684881 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:44Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.726466 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.726509 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.726536 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.726559 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.726571 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.828989 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.829032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.829048 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.829070 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.829093 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.931490 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.931566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.931579 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.931597 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:44 crc kubenswrapper[4688]: I1001 15:47:44.931625 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:44Z","lastTransitionTime":"2025-10-01T15:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.034195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.034237 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.034253 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.034268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.034278 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.079306 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.079511 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.079647 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.079653 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.079723 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:48:17.079704553 +0000 UTC m=+86.430344525 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.079756 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.079826 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:48:17.079805766 +0000 UTC m=+86.430445778 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.080052 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:17.080021892 +0000 UTC m=+86.430661864 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.137512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.137594 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.137614 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.137637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.137655 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.181112 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.181186 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181361 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181388 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181404 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181415 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181460 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181468 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:48:17.181451772 +0000 UTC m=+86.532091744 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181478 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.181586 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:48:17.181556195 +0000 UTC m=+86.532196177 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.239790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.239829 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.239838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.239852 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.239863 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.342230 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.342315 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.342333 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.342354 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.342371 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.380855 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.380855 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.381003 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.381077 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.380875 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:45 crc kubenswrapper[4688]: E1001 15:47:45.381145 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.444419 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.444494 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.444552 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.444585 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.444607 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.547777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.547841 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.547860 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.547883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.547900 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.651018 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.651076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.651094 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.651119 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.651141 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.754675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.754738 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.754754 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.754779 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.754796 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.857491 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.857596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.857615 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.858107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.858185 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.960801 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.960846 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.960862 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.960884 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:45 crc kubenswrapper[4688]: I1001 15:47:45.960899 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:45Z","lastTransitionTime":"2025-10-01T15:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.063839 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.063903 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.063921 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.063944 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.063961 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.171271 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.171324 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.171340 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.171372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.171388 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.275269 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.275450 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.275472 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.275583 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.275606 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.379829 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:46 crc kubenswrapper[4688]: E1001 15:47:46.379990 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.380027 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.380062 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.380079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.380098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.380114 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.482623 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.482657 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.482670 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.482689 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.482704 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.585836 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.585897 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.585914 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.585937 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.585954 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.688936 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.689003 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.689015 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.689030 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.689040 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.791240 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.791277 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.791286 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.791300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.791312 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.894943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.895004 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.895022 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.895046 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.895066 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.997540 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.997596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.997607 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.997625 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:46 crc kubenswrapper[4688]: I1001 15:47:46.997637 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:46Z","lastTransitionTime":"2025-10-01T15:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.099969 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.100180 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.100194 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.100206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.100214 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.203385 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.203442 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.203463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.203491 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.203510 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.306640 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.306698 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.306714 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.306741 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.306759 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.380909 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.381004 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:47 crc kubenswrapper[4688]: E1001 15:47:47.381082 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.381106 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:47 crc kubenswrapper[4688]: E1001 15:47:47.381277 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:47 crc kubenswrapper[4688]: E1001 15:47:47.381490 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.409879 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.409953 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.409976 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.410003 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.410020 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.512775 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.512822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.512833 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.512853 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.512865 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.615333 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.615388 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.615404 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.615426 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.615444 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.718713 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.718766 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.718783 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.718806 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.718824 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.821714 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.821761 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.821773 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.821790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.821802 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.924762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.924818 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.924837 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.924863 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:47 crc kubenswrapper[4688]: I1001 15:47:47.924879 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:47Z","lastTransitionTime":"2025-10-01T15:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.028107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.028171 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.028182 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.028199 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.028211 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.131741 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.131824 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.131849 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.131880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.131903 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.234458 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.234503 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.234514 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.234566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.234583 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.337977 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.338043 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.338063 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.338091 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.338108 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.380723 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:48 crc kubenswrapper[4688]: E1001 15:47:48.381121 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.442090 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.442401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.442652 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.442871 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.443072 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.545801 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.545883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.545898 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.545916 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.545930 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.648795 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.648857 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.648875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.648899 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.648917 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.751441 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.751493 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.751512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.751588 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.751648 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.856456 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.856976 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.857192 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.857438 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.857714 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.961359 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.961401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.961412 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.961428 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:48 crc kubenswrapper[4688]: I1001 15:47:48.961440 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:48Z","lastTransitionTime":"2025-10-01T15:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.064014 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.064073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.064092 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.064117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.064136 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.166411 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.166470 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.166489 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.166512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.166554 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.269695 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.269758 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.269779 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.269808 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.269829 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.373097 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.373182 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.373206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.373234 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.373258 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.380813 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.380849 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.380817 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.380979 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.381159 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.381254 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.476358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.476426 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.476449 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.476475 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.476497 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.580302 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.580395 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.580767 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.581096 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.581136 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.605112 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.628031 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.635183 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.635245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.635268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.635296 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.635313 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.648594 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.658911 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.665127 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.665172 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.665189 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.665211 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.665227 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.669212 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.687166 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.690808 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.693881 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.693923 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.693939 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.693960 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.693976 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.706985 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.714898 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.719713 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.719760 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.719777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.719800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.719817 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.729813 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.741473 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.750830 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.750896 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.750938 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.750971 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.750994 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.755637 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.773913 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: E1001 15:47:49.774142 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.776837 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.776906 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.776928 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.776958 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.776980 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.778703 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.809650 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.831769 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.850495 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.865206 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.879198 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.879255 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.879274 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.879297 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.879315 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.886241 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.906267 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.921755 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.936019 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.953159 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:49Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.982518 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.982656 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.982681 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.982713 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:49 crc kubenswrapper[4688]: I1001 15:47:49.982735 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:49Z","lastTransitionTime":"2025-10-01T15:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.085703 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.085777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.085819 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.085855 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.085879 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.189254 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.189325 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.189349 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.189380 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.189402 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.292623 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.292730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.292755 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.292787 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.292812 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.380681 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:50 crc kubenswrapper[4688]: E1001 15:47:50.380950 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.382026 4688 scope.go:117] "RemoveContainer" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" Oct 01 15:47:50 crc kubenswrapper[4688]: E1001 15:47:50.382295 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.395910 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.396011 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.396032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.396056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.396074 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.498830 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.499085 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.499174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.499243 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.499307 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.602676 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.602729 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.602771 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.602794 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.602808 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.706042 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.706382 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.706571 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.706729 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.706919 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.811417 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.811483 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.811506 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.811577 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.811603 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.914486 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.914584 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.914604 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.914630 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:50 crc kubenswrapper[4688]: I1001 15:47:50.914649 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:50Z","lastTransitionTime":"2025-10-01T15:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.016850 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.016897 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.016914 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.016938 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.016957 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.120865 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.120915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.120931 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.120957 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.120974 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.224265 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.224325 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.224342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.224367 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.224384 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.327978 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.328046 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.328338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.328373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.328395 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.379864 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:51 crc kubenswrapper[4688]: E1001 15:47:51.380044 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.380154 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:51 crc kubenswrapper[4688]: E1001 15:47:51.380265 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.380611 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:51 crc kubenswrapper[4688]: E1001 15:47:51.380701 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.404145 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.422016 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.431044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.431211 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.431235 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.431263 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.431284 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.442657 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.456929 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.470517 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.484094 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.499646 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.515551 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.528208 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.533828 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.533865 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.533876 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.533894 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.533908 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.542497 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.554823 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.565314 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.574331 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.589295 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.600596 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.616060 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.626309 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:51Z is after 2025-08-24T17:21:41Z" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.637076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.637117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.637130 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.637151 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.637163 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.740044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.740127 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.740143 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.740165 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.740180 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.845650 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.845741 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.845768 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.845800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.845824 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.949499 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.949596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.949619 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.949647 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:51 crc kubenswrapper[4688]: I1001 15:47:51.949668 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:51Z","lastTransitionTime":"2025-10-01T15:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.052424 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.052482 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.052615 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.052641 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.052660 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.155892 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.156370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.156389 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.156905 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.156968 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.260366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.260413 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.260429 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.260451 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.260467 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.362446 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.362501 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.362518 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.362595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.362617 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.379916 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:52 crc kubenswrapper[4688]: E1001 15:47:52.380080 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.466036 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.466132 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.466158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.466190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.466214 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.568755 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.568809 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.568827 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.568852 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.568869 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.671961 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.672039 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.672061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.672093 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.672117 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.775497 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.775573 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.775586 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.775608 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.775650 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.879110 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.879173 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.879192 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.879215 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.879232 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.982247 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.982310 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.982329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.982353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:52 crc kubenswrapper[4688]: I1001 15:47:52.982370 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:52Z","lastTransitionTime":"2025-10-01T15:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.085300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.085358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.085377 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.085401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.085422 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.188426 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.188479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.188498 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.188551 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.188572 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.291286 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.291361 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.291386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.291415 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.291438 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.380218 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.380291 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.380333 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:53 crc kubenswrapper[4688]: E1001 15:47:53.380430 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:53 crc kubenswrapper[4688]: E1001 15:47:53.380626 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:53 crc kubenswrapper[4688]: E1001 15:47:53.380764 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.393777 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.393840 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.393865 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.393899 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.393922 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.496730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.496783 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.496794 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.496813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.496824 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.600131 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.600195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.600219 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.600245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.600264 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.703643 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.703709 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.703732 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.703761 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.703784 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.807708 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.807775 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.807792 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.807815 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.807834 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.910164 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.910220 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.910238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.910260 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:53 crc kubenswrapper[4688]: I1001 15:47:53.910278 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:53Z","lastTransitionTime":"2025-10-01T15:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.013847 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.013938 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.013954 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.013978 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.013995 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.116791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.116852 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.116868 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.116891 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.116911 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.220352 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.220419 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.220437 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.220460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.220478 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.323307 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.323385 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.323411 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.323438 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.323455 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.380745 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:54 crc kubenswrapper[4688]: E1001 15:47:54.380934 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.425982 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.426051 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.426069 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.426093 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.426111 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.529647 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.529724 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.529743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.529780 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.529810 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.632416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.632498 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.632513 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.632561 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.632578 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.735328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.735366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.735378 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.735394 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.735405 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.837460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.837504 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.837515 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.837545 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.837561 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.939513 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.939553 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.939560 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.939573 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:54 crc kubenswrapper[4688]: I1001 15:47:54.939582 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:54Z","lastTransitionTime":"2025-10-01T15:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.042386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.042462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.042496 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.042557 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.042581 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.145032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.145077 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.145089 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.145107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.145119 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.247765 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.247833 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.247850 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.247873 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.247888 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.350341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.350404 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.350415 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.350430 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.350439 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.380714 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:55 crc kubenswrapper[4688]: E1001 15:47:55.380843 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.380986 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:55 crc kubenswrapper[4688]: E1001 15:47:55.381093 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.381007 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:55 crc kubenswrapper[4688]: E1001 15:47:55.381387 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.453984 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.454044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.454061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.454084 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.454103 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.556501 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.556617 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.556645 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.556704 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.556730 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.659390 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.659435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.659447 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.659467 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.659478 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.762183 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.762239 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.762256 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.762278 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.762295 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.864547 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.864579 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.864590 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.864604 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.864614 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.966851 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.966902 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.966917 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.966941 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:55 crc kubenswrapper[4688]: I1001 15:47:55.966956 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:55Z","lastTransitionTime":"2025-10-01T15:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.069169 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.069208 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.069220 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.069235 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.069246 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.171427 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.171461 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.171470 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.171482 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.171491 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.273385 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.273424 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.273432 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.273462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.273475 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.376080 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.376120 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.376132 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.376148 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.376158 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.380371 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:56 crc kubenswrapper[4688]: E1001 15:47:56.380577 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.478240 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.478292 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.478303 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.478320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.478331 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.580688 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.580731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.580740 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.580756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.580775 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.682635 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.682672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.682684 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.682697 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.682706 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.784875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.784912 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.784921 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.784937 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.784947 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.886995 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.887038 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.887050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.887065 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.887077 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.989999 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.990032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.990041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.990056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:56 crc kubenswrapper[4688]: I1001 15:47:56.990065 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:56Z","lastTransitionTime":"2025-10-01T15:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.092223 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.092291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.092309 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.092333 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.092351 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.194931 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.194975 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.194983 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.194997 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.195006 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.296967 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.297009 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.297017 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.297054 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.297063 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.380850 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.380947 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:57 crc kubenswrapper[4688]: E1001 15:47:57.380991 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.381013 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:57 crc kubenswrapper[4688]: E1001 15:47:57.381125 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:57 crc kubenswrapper[4688]: E1001 15:47:57.381178 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.399050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.399078 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.399087 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.399098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.399107 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.501388 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.501415 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.501423 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.501435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.501443 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.603619 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.603647 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.603655 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.603686 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.603695 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.705484 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.705577 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.705597 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.705618 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.705635 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.808236 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.808276 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.808291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.808312 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.808327 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.910479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.910534 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.910547 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.910561 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:57 crc kubenswrapper[4688]: I1001 15:47:57.910573 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:57Z","lastTransitionTime":"2025-10-01T15:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.012875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.012912 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.012921 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.012936 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.012946 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.115425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.115455 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.115466 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.115479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.115489 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.218419 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.218458 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.218467 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.218480 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.218491 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.320826 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.320851 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.320860 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.320871 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.320881 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.380599 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:58 crc kubenswrapper[4688]: E1001 15:47:58.380751 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.422595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.422621 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.422630 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.422642 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.422650 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.525174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.525206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.525216 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.525229 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.525238 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.627716 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.627760 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.627776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.627799 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.627816 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.730584 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.730622 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.730633 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.730649 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.730662 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.832984 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.833041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.833050 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.833065 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.833077 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.932437 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:47:58 crc kubenswrapper[4688]: E1001 15:47:58.932611 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:58 crc kubenswrapper[4688]: E1001 15:47:58.932721 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:48:30.932695771 +0000 UTC m=+100.283335763 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.936091 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.936142 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.936158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.936207 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:58 crc kubenswrapper[4688]: I1001 15:47:58.936233 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:58Z","lastTransitionTime":"2025-10-01T15:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.038161 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.038208 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.038226 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.038247 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.038264 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.141332 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.141372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.141381 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.141393 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.141402 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.244075 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.244111 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.244119 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.244134 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.244142 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.346834 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.346859 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.346868 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.346880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.346888 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.380430 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:47:59 crc kubenswrapper[4688]: E1001 15:47:59.380623 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.380667 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:47:59 crc kubenswrapper[4688]: E1001 15:47:59.380819 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.380880 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:47:59 crc kubenswrapper[4688]: E1001 15:47:59.380982 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.455109 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.455142 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.455154 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.455171 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.455183 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.557541 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.557598 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.557609 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.557624 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.557633 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.659999 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.660064 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.660081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.660106 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.660123 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.762813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.762844 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.762853 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.762867 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.762876 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.865293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.865329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.865338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.865353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.865364 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.967263 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.967338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.967355 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.967380 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:47:59 crc kubenswrapper[4688]: I1001 15:47:59.967399 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:47:59Z","lastTransitionTime":"2025-10-01T15:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.069286 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.069341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.069363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.069391 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.069414 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.070952 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.071008 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.071024 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.071047 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.071063 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.085718 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.088633 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.088659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.088669 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.088682 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.088690 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.102386 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.106837 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.106860 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.106870 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.106882 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.106905 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.118969 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.122497 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.122534 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.122542 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.122553 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.122568 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.132990 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.135968 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.135986 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.135994 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.136004 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.136012 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.146905 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.147009 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.171717 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.171807 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.171821 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.171836 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.171848 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.273848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.273926 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.273950 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.273974 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.273990 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.376043 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.376066 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.376076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.376090 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.376098 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.380445 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:00 crc kubenswrapper[4688]: E1001 15:48:00.380549 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.478276 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.478365 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.478385 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.478409 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.478427 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.581399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.581446 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.581460 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.581477 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.581490 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.683713 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.683759 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.683770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.683822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.683839 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.785718 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.785766 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.785781 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.785803 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.785834 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.841149 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/0.log" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.841210 4688 generic.go:334] "Generic (PLEG): container finished" podID="ac2218c2-479a-4efd-9c1b-ed0d3968b06b" containerID="329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342" exitCode=1 Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.841250 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerDied","Data":"329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.841718 4688 scope.go:117] "RemoveContainer" containerID="329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.858229 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.872974 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.881492 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.891479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.891537 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.891553 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.891567 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.891576 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.899181 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.914342 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.931111 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.950563 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.963895 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.975499 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.986351 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.997035 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.997061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.997070 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.997082 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:00 crc kubenswrapper[4688]: I1001 15:48:00.997091 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:00Z","lastTransitionTime":"2025-10-01T15:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.000771 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:00Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.013092 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.030214 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.043580 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.058924 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.071564 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.085810 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.099598 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.099645 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.099657 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.099675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.099688 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.202401 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.202452 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.202466 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.202487 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.202501 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.305366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.305415 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.305429 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.305450 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.305466 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.380109 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.380132 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.380171 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:01 crc kubenswrapper[4688]: E1001 15:48:01.380219 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:01 crc kubenswrapper[4688]: E1001 15:48:01.380381 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:01 crc kubenswrapper[4688]: E1001 15:48:01.381218 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.381704 4688 scope.go:117] "RemoveContainer" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.395639 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.410014 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.410040 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.410052 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.410065 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.410073 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.415738 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.428156 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.442671 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.451488 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.467117 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.476360 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.490387 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.501458 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.511310 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.512754 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.512792 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.512803 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.512815 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.512825 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.534007 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.547964 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.557729 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.569503 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.580675 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.591474 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.600850 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.615623 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.615664 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.615675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.615691 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.615706 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.717480 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.717859 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.717872 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.717889 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.717908 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.822546 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.822590 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.822601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.822617 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.822629 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.845776 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/2.log" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.847738 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.848160 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.849118 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/0.log" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.849150 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerStarted","Data":"78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.868737 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.879857 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.891692 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.911203 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.923073 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.924922 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.924950 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.925072 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.925087 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.925095 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:01Z","lastTransitionTime":"2025-10-01T15:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.935985 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.944295 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.956302 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.969473 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.982931 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:01 crc kubenswrapper[4688]: I1001 15:48:01.995075 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:01Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.005738 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.017813 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.027069 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.027108 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.027118 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.027131 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.027141 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.029193 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.039541 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.055472 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.065676 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.081277 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.123291 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.129462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.129489 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.129500 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.129514 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.129540 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.138446 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.161112 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.171227 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.179770 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.201415 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.211779 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.224339 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.232328 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.232372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.232384 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.232409 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.232421 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.237435 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.248597 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.257773 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.266811 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.274503 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.284646 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.294990 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.303549 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.338494 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.338621 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.338687 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.338751 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.338808 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.380371 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:02 crc kubenswrapper[4688]: E1001 15:48:02.380691 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.441303 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.441351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.441368 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.441390 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.441407 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.543762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.543800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.543813 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.543829 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.543840 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.646373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.646627 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.646874 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.647096 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.647225 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.749334 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.749368 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.749377 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.749389 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.749397 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.852231 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.852591 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.852757 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.852882 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.853009 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.854794 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/3.log" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.855746 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/2.log" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.858224 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" exitCode=1 Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.858309 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.858589 4688 scope.go:117] "RemoveContainer" containerID="3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.860435 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:48:02 crc kubenswrapper[4688]: E1001 15:48:02.861289 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.877984 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.892990 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.903769 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.933931 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dc073bb6833e5df7a40999b8b0b79d934c7457414f4833d1a10e20bf4bd757a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:47:37Z\\\",\\\"message\\\":\\\"e crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:47:37Z is after 2025-08-24T17:21:41Z]\\\\nI1001 15:47:37.302746 6276 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:02Z\\\",\\\"message\\\":\\\"316958 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1001 15:48:02.317045 6613 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1001 15:48:02.317063 6613 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-5sm44] creating logical port openshift-multus_network-metrics-daemon-5sm44 for pod on switch crc\\\\nI1001 15:48:02.317080 6613 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1001 15:48:02.317093 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.952295 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.960031 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.960075 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.960089 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.960106 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.960119 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:02Z","lastTransitionTime":"2025-10-01T15:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.973147 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:02 crc kubenswrapper[4688]: I1001 15:48:02.985791 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:02Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.004012 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.017308 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.031826 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.048166 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.062987 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.063044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.063061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.063112 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.063129 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.063508 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.082388 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.096828 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.107339 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.119224 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.131139 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.165484 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.165544 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.165557 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.165573 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.165584 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.267977 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.268198 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.268264 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.268333 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.268388 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.370783 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.370828 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.370840 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.370859 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.370872 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.380156 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.380209 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:03 crc kubenswrapper[4688]: E1001 15:48:03.380261 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:03 crc kubenswrapper[4688]: E1001 15:48:03.380336 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.380365 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:03 crc kubenswrapper[4688]: E1001 15:48:03.381687 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.473177 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.473238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.473257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.473281 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.473299 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.576104 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.576140 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.576152 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.576222 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.576258 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.678388 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.678428 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.678440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.678459 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.678470 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.780445 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.780494 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.780508 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.780540 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.780554 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.862317 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/3.log" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.864955 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:48:03 crc kubenswrapper[4688]: E1001 15:48:03.865108 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.878570 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.881847 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.881875 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.881883 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.881896 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.881905 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.891003 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.901404 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.911509 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.922294 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.930937 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.941033 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.951322 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.964151 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.973597 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.983981 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.984015 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.984026 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.984041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.984051 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:03Z","lastTransitionTime":"2025-10-01T15:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.985887 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:03 crc kubenswrapper[4688]: I1001 15:48:03.997477 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:03Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.013262 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:02Z\\\",\\\"message\\\":\\\"316958 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1001 15:48:02.317045 6613 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1001 15:48:02.317063 6613 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-5sm44] creating logical port openshift-multus_network-metrics-daemon-5sm44 for pod on switch crc\\\\nI1001 15:48:02.317080 6613 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1001 15:48:02.317093 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:04Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.023877 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:04Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.036230 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:04Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.045420 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:04Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.056056 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:04Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.086185 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.086258 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.086281 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.086313 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.086336 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.188672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.188715 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.188728 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.188744 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.188755 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.291170 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.291198 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.291206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.291218 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.291227 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.380391 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:04 crc kubenswrapper[4688]: E1001 15:48:04.380508 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.393317 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.393342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.393350 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.393362 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.393370 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.495670 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.495702 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.495709 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.495722 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.495731 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.598147 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.598206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.598225 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.598249 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.598267 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.700568 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.700601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.700608 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.700622 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.700631 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.802983 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.803012 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.803021 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.803035 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.803044 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.904987 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.905033 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.905051 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.905072 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:04 crc kubenswrapper[4688]: I1001 15:48:04.905088 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:04Z","lastTransitionTime":"2025-10-01T15:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.007272 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.007331 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.007341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.007356 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.007366 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.109886 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.109949 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.109963 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.109983 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.109997 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.212696 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.212753 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.212762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.212776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.212786 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.315311 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.315339 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.315349 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.315363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.315372 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.380464 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.380603 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:05 crc kubenswrapper[4688]: E1001 15:48:05.380691 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.380721 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:05 crc kubenswrapper[4688]: E1001 15:48:05.380865 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:05 crc kubenswrapper[4688]: E1001 15:48:05.380911 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.417609 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.417632 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.417641 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.417652 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.417660 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.520148 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.520181 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.520191 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.520205 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.520215 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.622514 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.622632 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.622647 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.622664 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.622676 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.725745 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.725842 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.725861 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.725916 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.725937 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.829365 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.829396 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.829404 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.829416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.829424 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.932232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.932293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.932309 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.932329 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:05 crc kubenswrapper[4688]: I1001 15:48:05.932345 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:05Z","lastTransitionTime":"2025-10-01T15:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.049158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.049213 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.049230 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.049251 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.049268 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.152480 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.152563 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.152580 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.152603 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.152621 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.255314 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.255360 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.255376 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.255398 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.255415 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.358784 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.358853 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.358877 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.358908 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.358929 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.380355 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:06 crc kubenswrapper[4688]: E1001 15:48:06.380646 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.461432 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.461471 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.461482 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.461498 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.461509 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.563933 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.564000 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.564017 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.564044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.564059 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.667254 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.667304 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.667321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.667344 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.667361 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.770495 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.770566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.770585 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.770610 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.770632 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.873759 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.873880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.873902 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.873926 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.873979 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.977431 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.977504 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.977563 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.977596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:06 crc kubenswrapper[4688]: I1001 15:48:06.977619 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:06Z","lastTransitionTime":"2025-10-01T15:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.080596 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.080656 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.080672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.080695 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.080715 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.183703 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.183771 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.183790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.183812 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.183828 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.286860 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.286917 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.286942 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.286970 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.286994 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.380476 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.380632 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:07 crc kubenswrapper[4688]: E1001 15:48:07.380801 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:07 crc kubenswrapper[4688]: E1001 15:48:07.381519 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.381985 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:07 crc kubenswrapper[4688]: E1001 15:48:07.382185 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.390923 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.391047 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.391067 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.391092 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.391111 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.493773 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.493897 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.493969 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.494006 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.494030 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.597386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.597461 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.597479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.597595 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.597624 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.700965 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.701033 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.701051 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.701079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.701098 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.804199 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.804260 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.804279 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.804306 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.804324 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.907557 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.907604 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.907617 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.907635 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:07 crc kubenswrapper[4688]: I1001 15:48:07.907648 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:07Z","lastTransitionTime":"2025-10-01T15:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.010911 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.010983 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.011002 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.011034 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.011052 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.114299 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.114362 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.114379 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.114406 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.114423 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.218152 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.218236 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.218255 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.218279 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.218296 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.321833 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.321912 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.321939 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.321971 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.321994 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.380031 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:08 crc kubenswrapper[4688]: E1001 15:48:08.380190 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.425149 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.425213 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.425225 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.425242 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.425252 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.528037 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.528104 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.528117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.528133 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.528146 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.631198 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.631261 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.631279 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.631305 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.631324 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.734955 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.735030 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.735052 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.735081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.735103 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.837565 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.837612 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.837623 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.837640 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.837654 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.940838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.940879 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.940890 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.940907 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:08 crc kubenswrapper[4688]: I1001 15:48:08.940918 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:08Z","lastTransitionTime":"2025-10-01T15:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.043857 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.043902 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.043919 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.043942 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.043958 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.147042 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.147107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.147124 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.147152 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.147170 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.249238 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.249316 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.249345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.249376 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.249430 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.352916 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.353053 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.353082 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.353111 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.353136 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.380492 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.380579 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:09 crc kubenswrapper[4688]: E1001 15:48:09.380692 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.380772 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:09 crc kubenswrapper[4688]: E1001 15:48:09.380938 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:09 crc kubenswrapper[4688]: E1001 15:48:09.381141 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.455920 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.456009 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.456030 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.456053 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.456102 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.559190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.559240 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.559257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.559278 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.559295 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.666098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.666158 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.666175 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.666200 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.666223 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.768821 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.768876 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.768895 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.768919 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.768938 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.871981 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.872087 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.872105 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.872130 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.872146 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.975284 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.975338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.975354 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.975377 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:09 crc kubenswrapper[4688]: I1001 15:48:09.975395 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:09Z","lastTransitionTime":"2025-10-01T15:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.078671 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.078730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.078747 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.078769 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.078790 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.182073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.182139 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.182162 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.182192 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.182214 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.260143 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.260218 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.260245 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.260274 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.260295 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.281098 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:10Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.286559 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.286613 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.286630 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.286653 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.286670 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.302930 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:10Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.307346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.307399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.307418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.307443 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.307460 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.323407 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:10Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.328479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.328562 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.328580 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.328604 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.328626 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.347874 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:10Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.360572 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.360623 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.360635 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.360651 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.360663 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.380086 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.380279 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.382014 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:10Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:10 crc kubenswrapper[4688]: E1001 15:48:10.382248 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.383998 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.384049 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.384070 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.384094 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.384112 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.396199 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.487437 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.487592 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.487620 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.487649 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.487670 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.590783 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.590877 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.590895 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.590919 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.590936 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.694321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.694382 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.694399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.694423 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.694441 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.797794 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.797879 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.797899 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.797925 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.797942 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.900821 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.900943 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.900963 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.900988 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:10 crc kubenswrapper[4688]: I1001 15:48:10.901005 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:10Z","lastTransitionTime":"2025-10-01T15:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.004470 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.004584 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.004605 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.004629 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.004645 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.107832 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.107897 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.107920 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.107951 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.107973 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.210618 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.210672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.210691 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.210715 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.210731 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.313754 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.313825 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.313843 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.313868 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.313884 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.380819 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.380898 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:11 crc kubenswrapper[4688]: E1001 15:48:11.380998 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.381063 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:11 crc kubenswrapper[4688]: E1001 15:48:11.381244 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:11 crc kubenswrapper[4688]: E1001 15:48:11.381574 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.399339 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.414497 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.417091 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.417249 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.417501 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.417638 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.417758 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.434678 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.454898 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.467841 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.483841 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.495199 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.506200 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.519032 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.521160 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.521189 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.521197 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.521210 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.521222 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.530441 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.542740 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67391ab2-4622-4310-9a58-f6cc60a35da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2931483443bdbbcf2d60d3d9e7f8e616a7a61c8ba9a05eb7d0b9534a4c3ba90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.568310 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.583300 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.597547 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.619884 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:02Z\\\",\\\"message\\\":\\\"316958 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1001 15:48:02.317045 6613 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1001 15:48:02.317063 6613 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-5sm44] creating logical port openshift-multus_network-metrics-daemon-5sm44 for pod on switch crc\\\\nI1001 15:48:02.317080 6613 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1001 15:48:02.317093 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.624068 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.624125 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.624139 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.624155 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.624165 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.637726 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.651406 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.661117 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:11Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.727017 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.727187 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.727214 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.727287 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.727356 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.830876 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.830941 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.830961 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.830986 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.831005 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.933678 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.933737 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.933753 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.933776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:11 crc kubenswrapper[4688]: I1001 15:48:11.933792 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:11Z","lastTransitionTime":"2025-10-01T15:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.036254 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.036301 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.036319 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.036341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.036358 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.139728 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.139800 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.139823 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.139854 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.139876 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.250298 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.250384 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.250410 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.250440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.250466 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.354903 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.355016 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.355047 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.355117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.355140 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.380514 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:12 crc kubenswrapper[4688]: E1001 15:48:12.380698 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.458263 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.458320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.458336 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.458358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.458375 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.561576 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.561637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.561656 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.561679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.561697 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.665448 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.665566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.665590 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.665617 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.665635 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.768700 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.768740 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.768753 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.768770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.768780 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.871312 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.871393 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.871411 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.871433 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.871452 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.974929 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.974985 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.975003 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.975028 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:12 crc kubenswrapper[4688]: I1001 15:48:12.975049 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:12Z","lastTransitionTime":"2025-10-01T15:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.078594 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.078659 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.078679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.078705 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.078723 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.181219 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.181297 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.181323 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.181347 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.181366 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.284507 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.284583 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.284601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.284628 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.284649 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.380584 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.380723 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:13 crc kubenswrapper[4688]: E1001 15:48:13.380873 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.380906 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:13 crc kubenswrapper[4688]: E1001 15:48:13.381083 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:13 crc kubenswrapper[4688]: E1001 15:48:13.381178 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.388915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.388970 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.388986 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.389008 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.389025 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.492620 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.492701 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.492724 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.492756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.492781 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.595684 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.595798 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.595825 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.595854 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.595878 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.698423 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.698482 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.698505 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.698558 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.698583 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.802153 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.802213 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.802230 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.802253 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.802270 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.904770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.904829 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.904847 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.904869 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:13 crc kubenswrapper[4688]: I1001 15:48:13.904889 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:13Z","lastTransitionTime":"2025-10-01T15:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.007857 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.007924 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.007941 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.007964 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.007979 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.110673 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.110773 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.110790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.110835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.110852 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.213098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.213137 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.213149 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.213164 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.213177 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.316261 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.316322 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.316339 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.316358 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.316373 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.379813 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:14 crc kubenswrapper[4688]: E1001 15:48:14.379935 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.419481 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.419568 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.419586 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.419609 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.419627 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.522981 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.523053 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.523070 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.523094 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.523113 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.626301 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.626743 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.626784 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.626820 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.626842 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.729563 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.729639 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.729663 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.729690 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.729714 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.832295 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.832345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.832355 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.832372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.832384 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.935820 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.935895 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.935918 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.935946 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:14 crc kubenswrapper[4688]: I1001 15:48:14.935968 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:14Z","lastTransitionTime":"2025-10-01T15:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.038807 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.038867 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.038882 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.038903 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.038919 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.141257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.141309 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.141324 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.141342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.141356 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.244073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.244174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.244188 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.244204 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.244215 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.347791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.347877 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.347907 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.347942 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.347964 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.380631 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.380731 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:15 crc kubenswrapper[4688]: E1001 15:48:15.380839 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.380857 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:15 crc kubenswrapper[4688]: E1001 15:48:15.381578 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:15 crc kubenswrapper[4688]: E1001 15:48:15.381742 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.382149 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:48:15 crc kubenswrapper[4688]: E1001 15:48:15.382439 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.451760 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.451820 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.451840 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.451870 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.451890 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.556178 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.556234 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.556256 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.556288 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.556309 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.659052 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.659113 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.659132 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.659155 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.659172 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.762418 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.762479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.762497 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.762519 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.762581 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.865469 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.865628 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.865656 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.865687 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.865712 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.968836 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.968913 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.968933 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.968959 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:15 crc kubenswrapper[4688]: I1001 15:48:15.968978 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:15Z","lastTransitionTime":"2025-10-01T15:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.072225 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.072298 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.072321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.072351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.072377 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.175477 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.175540 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.175550 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.175564 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.175575 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.278151 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.278190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.278202 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.278218 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.278229 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.380423 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:16 crc kubenswrapper[4688]: E1001 15:48:16.380615 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.381002 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.381037 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.381052 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.381074 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.381092 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.483462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.483540 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.483557 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.483574 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.483586 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.586742 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.587103 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.587345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.587552 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.587764 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.690367 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.690439 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.690483 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.690512 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.690561 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.793365 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.793427 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.793453 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.793482 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.793502 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.895829 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.895886 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.895900 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.895920 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.895932 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.999278 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.999323 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.999334 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.999351 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:16 crc kubenswrapper[4688]: I1001 15:48:16.999362 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:16Z","lastTransitionTime":"2025-10-01T15:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.102463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.102570 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.102590 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.102614 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.102630 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.135006 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.135202 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.135303 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.135398 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:21.135359876 +0000 UTC m=+150.485999878 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.135419 4688 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.135402 4688 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.135497 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:49:21.135472619 +0000 UTC m=+150.486112611 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.135578 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 15:49:21.135511041 +0000 UTC m=+150.486151033 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.205342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.205408 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.205425 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.205450 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.205467 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.235856 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.235958 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236066 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236090 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236106 4688 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236105 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236129 4688 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236143 4688 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236164 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 15:49:21.23614722 +0000 UTC m=+150.586787192 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.236189 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 15:49:21.236172721 +0000 UTC m=+150.586812693 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.309028 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.309073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.309088 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.309107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.309122 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.380384 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.380479 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.380482 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.380627 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.380937 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:17 crc kubenswrapper[4688]: E1001 15:48:17.381009 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.412694 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.412755 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.412772 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.412796 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.412812 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.515671 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.515719 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.515730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.515748 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.515760 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.623120 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.623168 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.623178 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.623196 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.623209 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.726933 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.727280 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.727480 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.727710 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.727860 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.830670 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.830744 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.830762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.830790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.830812 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.934330 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.934414 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.934440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.934476 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:17 crc kubenswrapper[4688]: I1001 15:48:17.934503 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:17Z","lastTransitionTime":"2025-10-01T15:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.036661 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.036704 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.036715 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.036731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.036745 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.140356 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.140440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.140459 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.140483 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.140501 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.243923 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.243974 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.243991 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.244014 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.244032 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.346671 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.346738 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.346762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.346790 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.346811 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.380711 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:18 crc kubenswrapper[4688]: E1001 15:48:18.380930 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.450063 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.450130 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.450149 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.450174 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.450192 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.553399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.553453 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.553475 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.553503 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.553557 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.657021 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.657068 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.657086 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.657114 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.657133 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.759633 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.759691 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.759708 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.759731 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.759749 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.861966 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.862007 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.862018 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.862033 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.862043 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.964651 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.964720 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.964739 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.964764 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:18 crc kubenswrapper[4688]: I1001 15:48:18.964781 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:18Z","lastTransitionTime":"2025-10-01T15:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.067882 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.067937 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.067955 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.067978 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.067996 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.170586 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.170628 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.170639 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.170654 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.170667 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.273807 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.273870 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.273887 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.273911 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.273928 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.377293 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.377348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.377373 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.377414 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.377439 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.381220 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.381322 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.381236 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:19 crc kubenswrapper[4688]: E1001 15:48:19.381513 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:19 crc kubenswrapper[4688]: E1001 15:48:19.381730 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:19 crc kubenswrapper[4688]: E1001 15:48:19.381904 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.480663 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.480739 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.480762 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.480791 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.480813 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.583383 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.583459 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.583483 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.583510 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.583568 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.686457 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.686547 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.686566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.686589 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.686607 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.789736 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.789825 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.789842 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.789867 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.789885 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.893799 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.893929 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.893955 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.894026 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.894052 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.997348 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.997454 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.997479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.997614 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:19 crc kubenswrapper[4688]: I1001 15:48:19.997642 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:19Z","lastTransitionTime":"2025-10-01T15:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.101488 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.101576 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.101597 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.101620 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.101637 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.205044 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.205157 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.205195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.205232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.205257 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.309209 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.309287 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.309310 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.309338 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.309363 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.380225 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.380434 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.412223 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.412291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.412311 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.412336 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.412355 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.413973 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.414035 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.414054 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.414076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.414094 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.434366 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.439250 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.439306 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.439323 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.439346 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.439363 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.459021 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.463704 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.463756 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.463776 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.463803 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.463822 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.484452 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.489774 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.489871 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.489899 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.489974 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.490000 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.511936 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.517032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.517088 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.517105 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.517130 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.517146 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.539227 4688 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"71eff274-2d68-4a0d-9181-6c8c42ab6723\\\",\\\"systemUUID\\\":\\\"09adb3c3-2d03-448f-a3d6-41cc884a6e83\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:20Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:20 crc kubenswrapper[4688]: E1001 15:48:20.539599 4688 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.542030 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.542116 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.542138 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.542196 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.542226 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.645225 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.645283 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.645300 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.645327 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.645351 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.748874 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.748958 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.748987 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.749019 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.749042 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.852778 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.852826 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.852858 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.852880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.852894 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.955172 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.955215 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.955231 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.955252 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:20 crc kubenswrapper[4688]: I1001 15:48:20.955268 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:20Z","lastTransitionTime":"2025-10-01T15:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.058450 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.058508 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.058565 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.058597 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.058621 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.162071 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.162138 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.162161 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.162190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.162215 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.265144 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.265201 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.265217 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.265240 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.265256 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.368620 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.369121 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.369326 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.369581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.369803 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.380034 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:21 crc kubenswrapper[4688]: E1001 15:48:21.380194 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.380401 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.380686 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:21 crc kubenswrapper[4688]: E1001 15:48:21.380912 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:21 crc kubenswrapper[4688]: E1001 15:48:21.381163 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.413881 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7628fc05-465a-425f-b13a-995f8d5fa914\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:02Z\\\",\\\"message\\\":\\\"316958 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1001 15:48:02.317045 6613 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1001 15:48:02.317063 6613 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-5sm44] creating logical port openshift-multus_network-metrics-daemon-5sm44 for pod on switch crc\\\\nI1001 15:48:02.317080 6613 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1001 15:48:02.317093 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:48:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c8bsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xgf9s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.437681 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcfs4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac2218c2-479a-4efd-9c1b-ed0d3968b06b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T15:48:00Z\\\",\\\"message\\\":\\\"2025-10-01T15:47:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa\\\\n2025-10-01T15:47:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5aaf5872-07d3-4fab-bef9-c78a97d8e7aa to /host/opt/cni/bin/\\\\n2025-10-01T15:47:15Z [verbose] multus-daemon started\\\\n2025-10-01T15:47:15Z [verbose] Readiness Indicator file check\\\\n2025-10-01T15:48:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcfs4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.464421 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e83f8b9c-424d-4983-b5d3-83b52939c5eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1908355d5db12e2554d7b02d677598d5888aabaf7ce7eaf3b72aea4642d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd065ed1184348291f5ba7b225ca792c489d64ab41d94bdc68cdd3558520d83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706f5eda567067ca5fc2eef7f5c8400c0909ebbb47dca404d68dc64d1d1bc55b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05b7ca5d9595e123474a27d18c6c10b24d3b891ef2f4963123b874355f39f7d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3058ba1f4cf28e373c5aa3f7bdf25f0c0a3884eddb807fbf828d6bd5c82d4aaf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39608446a70846167adceb1e86f6114406f92b7647f3f3c405bf53671e66fbda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ddd956582a58ef549faf64a5010a36ce91cee9303231317e800bd5cdad6f14a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6wrpp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nrzjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.472412 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.472486 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.472515 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.472588 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.472625 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.483902 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xq8gs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6900ee15-3789-4e35-8d10-573acac2f5bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba7396b3c47c50254e4f6e9c37b3d0ca36209d208253ebb253a3c7dada282a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88k4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xq8gs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.500954 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67391ab2-4622-4310-9a58-f6cc60a35da5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2931483443bdbbcf2d60d3d9e7f8e616a7a61c8ba9a05eb7d0b9534a4c3ba90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c44c1484910b88ff1db68f1caa533f1e1998cc49fa2bcbfc3c5bc4025a8bf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.522579 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a86a341-f518-4908-abf0-9c7143787692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20799e8cd13f114de9dbdce4eb9aaa14bb756a887ae7b2469cf4025fc0479d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01e821210c6b8a6ddd049780366488dcd5e4755eb27e469008db8b57338e2ee0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://317ed6bd617fe535aa2c4305536bfd95f0e1956dff3819c49a76baa3d3b79079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f444bfe7afb7393d8d0dca1d95afa42a524bb397cde673972c00d1ca1029d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4b33c056050420f709b3d7441eaeb9dae8c7a861b25121ee3e21b4a8063721\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T15:47:12Z\\\",\\\"message\\\":\\\"le observer\\\\nW1001 15:47:12.354485 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1001 15:47:12.354742 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 15:47:12.356001 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1921256202/tls.crt::/tmp/serving-cert-1921256202/tls.key\\\\\\\"\\\\nI1001 15:47:12.746133 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 15:47:12.751967 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 15:47:12.751995 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 15:47:12.752030 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 15:47:12.752040 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 15:47:12.758345 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1001 15:47:12.758376 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758382 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 15:47:12.758388 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 15:47:12.758393 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 15:47:12.758397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 15:47:12.758400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1001 15:47:12.758402 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1001 15:47:12.760493 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b0eb81bca42985a934b557f29537eebf0487f8c0adcf4179bf07cc8757e4d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3ab9091ce9c5c93d84e1d6f646693f34ff82fa7ff5bc37b983691e8aa915d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.541213 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"619783c9-56d6-4e76-96a7-2b59f4e6a466\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8985bf0d8fc40e82ee4c0b2baf04700529c0490a95c2f15f571f2659db7c161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd300a3c2e3352e5886a1d01583f177842439c26d3d3a2c7b09358a9bcf47f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc1511f39143b27af203eda9b50ec57599849027768d8f73721a4c1a00d12344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b9f093e971a2020adc2cead25955333680586b0f0c6d7713db6da47b72e755cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T15:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.559017 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.576370 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.576479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.576566 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.576600 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.576618 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.577724 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5sm44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c987bdd3-e8bc-473e-99b2-61a143a95543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78gwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:27Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5sm44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.598912 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"532709fb-1e8e-407f-ba42-e68235ab6029\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44ff63f1090bb88410db2928dbd2b80934896d2166df04a731b34c9a36f11f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c7969faea7edc9b77a58db49c769be92789551f8b2a5d10566a87b59aa816e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d564e1de2372020df8985df05beabbbcbf9106bf277ae72cb4367fbd3bdd7d8b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6faf39ffb65f25bcb33273fdc1320d80b1311686b5cc1a6b332671b264f605d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:46:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.619144 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.635176 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fd7067-ac62-426d-b3a3-dca2750d14db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4814ead7b452e3ad87dfa7ea81a0a5f49a584a3c7bc98051f62997b9c13bb181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lq7cn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6ndlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.651462 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c2b8d9c-d29c-497d-bac1-3eba88343919\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caeda02390173a17f5292783fdafb71c19e1c8b74f99bc904aa04fdb5823facf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fc2beefda3afb44f708dd9c6af8debea0a70e301d81164bd055d113c12e635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55lj6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-trwg6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.670391 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f3ef0bfa2c348745c9081226bd5d61e3a49f2913b386d884d9798809c264110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.680172 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.680214 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.680250 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.680269 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.680283 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.688110 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd457e4151abd91829e8223875f40b9abe64180e2e6c9bbc7ee07ad8e0862003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.707984 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.727445 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6efa7c15d5d5b97667ef5d9e1309b9caa5d381caa9e33b26ff3b944e69a02ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f19e2e98fa3dbf862f50aab825b959c4aa06f046c2fc6d0607ca78d9c15e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.743464 4688 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ccw98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"94ed6916-e5b4-40b9-87ed-fee49ecb1c28\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T15:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca74445585bc9cfc0bfd65c4194c8d783c8fcc30a643f146531b804a9f9997ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T15:47:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9pt8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T15:47:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ccw98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T15:48:21Z is after 2025-08-24T17:21:41Z" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.782184 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.782232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.782248 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.782268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.782281 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.885516 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.885581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.885597 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.885617 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.885632 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.988061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.988097 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.988107 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.988119 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:21 crc kubenswrapper[4688]: I1001 15:48:21.988128 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:21Z","lastTransitionTime":"2025-10-01T15:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.091041 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.091081 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.091093 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.091109 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.091121 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.194165 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.194250 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.194267 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.194291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.194310 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.296863 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.296915 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.296931 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.296955 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.296972 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.380694 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:22 crc kubenswrapper[4688]: E1001 15:48:22.381240 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.402046 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.402099 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.402115 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.402138 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.402156 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.506080 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.506126 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.506144 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.506168 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.506186 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.609438 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.609495 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.609511 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.609562 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.609579 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.713143 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.713308 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.713336 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.713435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.713488 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.817200 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.817257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.817283 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.817311 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.817330 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.920298 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.920363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.920384 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.920410 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:22 crc kubenswrapper[4688]: I1001 15:48:22.920428 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:22Z","lastTransitionTime":"2025-10-01T15:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.023250 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.023345 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.023363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.023385 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.023403 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.126372 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.126434 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.126453 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.126475 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.126492 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.229758 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.229830 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.229856 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.229890 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.229918 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.332132 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.332189 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.332212 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.332239 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.332262 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.380986 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:23 crc kubenswrapper[4688]: E1001 15:48:23.381133 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.381189 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.380989 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:23 crc kubenswrapper[4688]: E1001 15:48:23.381313 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:23 crc kubenswrapper[4688]: E1001 15:48:23.381442 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.435601 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.435660 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.435678 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.435702 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.435719 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.538769 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.538822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.538844 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.538873 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.538894 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.642581 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.642637 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.642653 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.642679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.642696 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.746190 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.746258 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.746277 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.746301 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.746319 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.849973 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.850057 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.850079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.850109 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.850131 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.953325 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.953387 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.953408 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.953436 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:23 crc kubenswrapper[4688]: I1001 15:48:23.953459 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:23Z","lastTransitionTime":"2025-10-01T15:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.056362 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.056419 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.056435 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.056462 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.056485 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.159588 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.159654 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.159675 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.159704 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.159721 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.262661 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.262770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.262796 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.262818 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.262835 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.366241 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.366279 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.366289 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.366303 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.366312 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.380164 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:24 crc kubenswrapper[4688]: E1001 15:48:24.380569 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.469082 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.469139 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.469161 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.469193 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.469213 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.572837 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.572917 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.572941 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.572967 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.572984 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.676183 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.676251 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.676268 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.676291 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.676315 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.779142 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.779674 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.779903 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.780160 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.780494 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.883736 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.883804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.883827 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.883855 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.883876 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.986954 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.987024 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.987043 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.987069 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:24 crc kubenswrapper[4688]: I1001 15:48:24.987088 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:24Z","lastTransitionTime":"2025-10-01T15:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.089077 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.089559 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.089719 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.089886 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.090018 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.193492 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.193591 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.193610 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.193634 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.193651 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.297202 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.297639 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.297804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.298012 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.298156 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.380489 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.380623 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.380644 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:25 crc kubenswrapper[4688]: E1001 15:48:25.381213 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:25 crc kubenswrapper[4688]: E1001 15:48:25.381372 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:25 crc kubenswrapper[4688]: E1001 15:48:25.381700 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.401880 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.401956 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.401980 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.402016 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.402038 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.505519 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.505610 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.505622 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.505645 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.505662 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.608668 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.608738 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.608760 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.608789 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.608812 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.712121 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.712186 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.712204 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.712228 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.712248 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.815831 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.815893 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.815909 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.815932 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.815948 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.919326 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.919388 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.919407 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.919429 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:25 crc kubenswrapper[4688]: I1001 15:48:25.919446 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:25Z","lastTransitionTime":"2025-10-01T15:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.022688 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.022988 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.023017 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.023046 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.023069 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.126736 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.126804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.126822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.126848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.126867 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.230214 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.230321 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.230342 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.230364 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.230380 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.333297 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.333353 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.333365 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.333386 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.333400 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.379986 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:26 crc kubenswrapper[4688]: E1001 15:48:26.380185 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.436730 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.436804 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.436822 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.436844 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.436861 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.539995 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.540056 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.540073 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.540096 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.540113 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.643252 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.643332 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.643369 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.643399 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.643420 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.745779 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.745864 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.745888 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.745922 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.745949 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.849140 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.849214 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.849232 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.849257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.849276 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.952187 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.952261 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.952281 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.952307 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:26 crc kubenswrapper[4688]: I1001 15:48:26.952326 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:26Z","lastTransitionTime":"2025-10-01T15:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.055797 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.055870 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.055891 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.055916 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.055938 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.159010 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.159076 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.159098 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.159126 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.159148 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.263360 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.263440 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.263463 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.263491 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.263544 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.366671 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.366736 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.366752 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.366775 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.366825 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.379925 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:27 crc kubenswrapper[4688]: E1001 15:48:27.380111 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.380451 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:27 crc kubenswrapper[4688]: E1001 15:48:27.380647 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.380770 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:27 crc kubenswrapper[4688]: E1001 15:48:27.380909 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.469436 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.469507 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.469552 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.469574 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.469589 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.573131 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.573185 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.573202 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.573223 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.573240 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.676479 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.676554 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.676570 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.676593 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.676610 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.779635 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.779695 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.779713 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.779737 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.779760 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.884095 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.884154 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.884176 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.884204 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.884227 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.987664 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.987726 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.987745 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.987770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:27 crc kubenswrapper[4688]: I1001 15:48:27.987788 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:27Z","lastTransitionTime":"2025-10-01T15:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.090155 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.090195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.090205 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.090218 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.090226 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.194115 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.194195 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.194219 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.194252 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.194274 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.297809 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.297902 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.297924 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.297947 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.297963 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.380612 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:28 crc kubenswrapper[4688]: E1001 15:48:28.380781 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.401011 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.401069 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.401092 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.401117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.401138 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.504151 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.504206 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.504222 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.504244 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.504261 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.613177 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.613231 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.613250 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.613274 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.613290 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.715983 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.716048 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.716061 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.716079 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.716093 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.819848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.820005 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.820032 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.820101 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.820125 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.923187 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.923257 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.923368 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.923455 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:28 crc kubenswrapper[4688]: I1001 15:48:28.923486 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:28Z","lastTransitionTime":"2025-10-01T15:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.027148 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.027233 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.027258 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.027286 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.027309 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.130760 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.130816 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.130835 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.130859 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.130876 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.234117 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.234320 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.234341 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.234366 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.234386 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.337726 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.337784 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.337801 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.337824 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.337841 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.381030 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:29 crc kubenswrapper[4688]: E1001 15:48:29.381189 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.381036 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:29 crc kubenswrapper[4688]: E1001 15:48:29.381671 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.381716 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:29 crc kubenswrapper[4688]: E1001 15:48:29.381961 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.382240 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:48:29 crc kubenswrapper[4688]: E1001 15:48:29.382478 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xgf9s_openshift-ovn-kubernetes(7628fc05-465a-425f-b13a-995f8d5fa914)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.440281 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.440325 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.440344 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.440371 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.440389 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.543768 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.544679 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.544739 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.544770 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.544814 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.648170 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.648261 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.648281 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.648335 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.648353 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.751202 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.751270 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.751290 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.751315 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.751334 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.854217 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.854696 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.854716 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.854740 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.854756 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.957359 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.957400 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.957416 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.957439 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:29 crc kubenswrapper[4688]: I1001 15:48:29.957455 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:29Z","lastTransitionTime":"2025-10-01T15:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.060848 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.060908 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.060924 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.061003 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.061021 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.164280 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.164343 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.164363 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.164390 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.164409 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.267445 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.267509 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.267574 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.267607 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.267630 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.370412 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.370450 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.370466 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.370486 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.370501 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.380628 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:30 crc kubenswrapper[4688]: E1001 15:48:30.380811 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.473603 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.473672 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.473696 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.473726 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.473748 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.576774 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.577185 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.577374 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.577612 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.577804 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.654470 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.654778 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.654838 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.654868 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.654890 4688 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T15:48:30Z","lastTransitionTime":"2025-10-01T15:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.722406 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf"] Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.723011 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.726409 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.726773 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.727060 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.727450 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.776431 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd9fd115-0afb-42d8-afdc-ece27c8f2227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.776517 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd9fd115-0afb-42d8-afdc-ece27c8f2227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.776731 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.776844 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fd9fd115-0afb-42d8-afdc-ece27c8f2227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.776913 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.836794 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ccw98" podStartSLOduration=77.836756173 podStartE2EDuration="1m17.836756173s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.836397163 +0000 UTC m=+100.187037165" watchObservedRunningTime="2025-10-01 15:48:30.836756173 +0000 UTC m=+100.187396145" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878424 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fd9fd115-0afb-42d8-afdc-ece27c8f2227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878501 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878636 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878667 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd9fd115-0afb-42d8-afdc-ece27c8f2227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878755 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd9fd115-0afb-42d8-afdc-ece27c8f2227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878898 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.878976 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/fd9fd115-0afb-42d8-afdc-ece27c8f2227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.881511 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fd9fd115-0afb-42d8-afdc-ece27c8f2227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.901799 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd9fd115-0afb-42d8-afdc-ece27c8f2227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.903036 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jcfs4" podStartSLOduration=78.903020486 podStartE2EDuration="1m18.903020486s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.902283644 +0000 UTC m=+100.252923616" watchObservedRunningTime="2025-10-01 15:48:30.903020486 +0000 UTC m=+100.253660458" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.907386 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd9fd115-0afb-42d8-afdc-ece27c8f2227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lqltf\" (UID: \"fd9fd115-0afb-42d8-afdc-ece27c8f2227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.939656 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-nrzjb" podStartSLOduration=78.939637499 podStartE2EDuration="1m18.939637499s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.927279707 +0000 UTC m=+100.277919699" watchObservedRunningTime="2025-10-01 15:48:30.939637499 +0000 UTC m=+100.290277471" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.953708 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xq8gs" podStartSLOduration=78.9536811 podStartE2EDuration="1m18.9536811s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.940667059 +0000 UTC m=+100.291307061" watchObservedRunningTime="2025-10-01 15:48:30.9536811 +0000 UTC m=+100.304321092" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.954826 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.954817034 podStartE2EDuration="20.954817034s" podCreationTimestamp="2025-10-01 15:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.953613358 +0000 UTC m=+100.304253330" watchObservedRunningTime="2025-10-01 15:48:30.954817034 +0000 UTC m=+100.305457036" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.977350 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.977331394 podStartE2EDuration="1m17.977331394s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.97617621 +0000 UTC m=+100.326816182" watchObservedRunningTime="2025-10-01 15:48:30.977331394 +0000 UTC m=+100.327971366" Oct 01 15:48:30 crc kubenswrapper[4688]: I1001 15:48:30.979776 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:30 crc kubenswrapper[4688]: E1001 15:48:30.979901 4688 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:48:30 crc kubenswrapper[4688]: E1001 15:48:30.979965 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs podName:c987bdd3-e8bc-473e-99b2-61a143a95543 nodeName:}" failed. No retries permitted until 2025-10-01 15:49:34.97994539 +0000 UTC m=+164.330585352 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs") pod "network-metrics-daemon-5sm44" (UID: "c987bdd3-e8bc-473e-99b2-61a143a95543") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.011616 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.011591868 podStartE2EDuration="47.011591868s" podCreationTimestamp="2025-10-01 15:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:30.99393892 +0000 UTC m=+100.344578922" watchObservedRunningTime="2025-10-01 15:48:31.011591868 +0000 UTC m=+100.362231860" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.045644 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.045620595 podStartE2EDuration="1m17.045620595s" podCreationTimestamp="2025-10-01 15:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:31.044978326 +0000 UTC m=+100.395618318" watchObservedRunningTime="2025-10-01 15:48:31.045620595 +0000 UTC m=+100.396260597" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.048283 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.090836 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podStartSLOduration=79.0908173 podStartE2EDuration="1m19.0908173s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:31.088458281 +0000 UTC m=+100.439098243" watchObservedRunningTime="2025-10-01 15:48:31.0908173 +0000 UTC m=+100.441457272" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.102627 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-trwg6" podStartSLOduration=78.102609256 podStartE2EDuration="1m18.102609256s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:31.102172073 +0000 UTC m=+100.452812045" watchObservedRunningTime="2025-10-01 15:48:31.102609256 +0000 UTC m=+100.453249228" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.379908 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.380041 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.380936 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:31 crc kubenswrapper[4688]: E1001 15:48:31.380967 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:31 crc kubenswrapper[4688]: E1001 15:48:31.381050 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:31 crc kubenswrapper[4688]: E1001 15:48:31.381199 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.974220 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" event={"ID":"fd9fd115-0afb-42d8-afdc-ece27c8f2227","Type":"ContainerStarted","Data":"79b33890c38ac1eab5f6d1e24b828cf99b34a8e0baf1d7f5355b42f37b2d4654"} Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.974299 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" event={"ID":"fd9fd115-0afb-42d8-afdc-ece27c8f2227","Type":"ContainerStarted","Data":"8610b5a445282878390bf8f3bbaec4933958d638bda53ad644b6344201f7f97f"} Oct 01 15:48:31 crc kubenswrapper[4688]: I1001 15:48:31.993811 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lqltf" podStartSLOduration=79.993790396 podStartE2EDuration="1m19.993790396s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:31.992664103 +0000 UTC m=+101.343304085" watchObservedRunningTime="2025-10-01 15:48:31.993790396 +0000 UTC m=+101.344430378" Oct 01 15:48:32 crc kubenswrapper[4688]: I1001 15:48:32.379833 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:32 crc kubenswrapper[4688]: E1001 15:48:32.379996 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:33 crc kubenswrapper[4688]: I1001 15:48:33.379901 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:33 crc kubenswrapper[4688]: I1001 15:48:33.379975 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:33 crc kubenswrapper[4688]: I1001 15:48:33.379913 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:33 crc kubenswrapper[4688]: E1001 15:48:33.380070 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:33 crc kubenswrapper[4688]: E1001 15:48:33.380169 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:33 crc kubenswrapper[4688]: E1001 15:48:33.380378 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:34 crc kubenswrapper[4688]: I1001 15:48:34.380509 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:34 crc kubenswrapper[4688]: E1001 15:48:34.380727 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:35 crc kubenswrapper[4688]: I1001 15:48:35.380833 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:35 crc kubenswrapper[4688]: I1001 15:48:35.380918 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:35 crc kubenswrapper[4688]: I1001 15:48:35.381097 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:35 crc kubenswrapper[4688]: E1001 15:48:35.382344 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:35 crc kubenswrapper[4688]: E1001 15:48:35.382462 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:35 crc kubenswrapper[4688]: E1001 15:48:35.382118 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:36 crc kubenswrapper[4688]: I1001 15:48:36.380752 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:36 crc kubenswrapper[4688]: E1001 15:48:36.380953 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:37 crc kubenswrapper[4688]: I1001 15:48:37.386886 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:37 crc kubenswrapper[4688]: I1001 15:48:37.386914 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:37 crc kubenswrapper[4688]: I1001 15:48:37.386921 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:37 crc kubenswrapper[4688]: E1001 15:48:37.387089 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:37 crc kubenswrapper[4688]: E1001 15:48:37.387205 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:37 crc kubenswrapper[4688]: E1001 15:48:37.387317 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:38 crc kubenswrapper[4688]: I1001 15:48:38.380945 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:38 crc kubenswrapper[4688]: E1001 15:48:38.381160 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:39 crc kubenswrapper[4688]: I1001 15:48:39.380926 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:39 crc kubenswrapper[4688]: E1001 15:48:39.381143 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:39 crc kubenswrapper[4688]: I1001 15:48:39.381453 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:39 crc kubenswrapper[4688]: E1001 15:48:39.381595 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:39 crc kubenswrapper[4688]: I1001 15:48:39.381793 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:39 crc kubenswrapper[4688]: E1001 15:48:39.381955 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:40 crc kubenswrapper[4688]: I1001 15:48:40.379791 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:40 crc kubenswrapper[4688]: E1001 15:48:40.379897 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:41 crc kubenswrapper[4688]: I1001 15:48:41.380459 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:41 crc kubenswrapper[4688]: I1001 15:48:41.382114 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:41 crc kubenswrapper[4688]: I1001 15:48:41.382146 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:41 crc kubenswrapper[4688]: E1001 15:48:41.382401 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:41 crc kubenswrapper[4688]: E1001 15:48:41.382695 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:41 crc kubenswrapper[4688]: E1001 15:48:41.382765 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:41 crc kubenswrapper[4688]: I1001 15:48:41.409464 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 01 15:48:42 crc kubenswrapper[4688]: I1001 15:48:42.380731 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:42 crc kubenswrapper[4688]: E1001 15:48:42.381199 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:43 crc kubenswrapper[4688]: I1001 15:48:43.380260 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:43 crc kubenswrapper[4688]: I1001 15:48:43.380276 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:43 crc kubenswrapper[4688]: E1001 15:48:43.381891 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:43 crc kubenswrapper[4688]: I1001 15:48:43.380297 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:43 crc kubenswrapper[4688]: E1001 15:48:43.382055 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:43 crc kubenswrapper[4688]: E1001 15:48:43.382247 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:44 crc kubenswrapper[4688]: I1001 15:48:44.382100 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:44 crc kubenswrapper[4688]: E1001 15:48:44.382252 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:44 crc kubenswrapper[4688]: I1001 15:48:44.382687 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.018897 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/3.log" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.022130 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerStarted","Data":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.022628 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.060884 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=4.060870179 podStartE2EDuration="4.060870179s" podCreationTimestamp="2025-10-01 15:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:45.060350334 +0000 UTC m=+114.410990286" watchObservedRunningTime="2025-10-01 15:48:45.060870179 +0000 UTC m=+114.411510151" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.087725 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podStartSLOduration=93.087710106 podStartE2EDuration="1m33.087710106s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:45.086632084 +0000 UTC m=+114.437272056" watchObservedRunningTime="2025-10-01 15:48:45.087710106 +0000 UTC m=+114.438350058" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.401137 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.401214 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:45 crc kubenswrapper[4688]: E1001 15:48:45.401336 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.401079 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:45 crc kubenswrapper[4688]: E1001 15:48:45.401467 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:45 crc kubenswrapper[4688]: E1001 15:48:45.401478 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.419861 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5sm44"] Oct 01 15:48:45 crc kubenswrapper[4688]: I1001 15:48:45.419979 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:45 crc kubenswrapper[4688]: E1001 15:48:45.420100 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:47 crc kubenswrapper[4688]: I1001 15:48:47.380623 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:47 crc kubenswrapper[4688]: I1001 15:48:47.380729 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:47 crc kubenswrapper[4688]: I1001 15:48:47.380738 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:47 crc kubenswrapper[4688]: I1001 15:48:47.380766 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:47 crc kubenswrapper[4688]: E1001 15:48:47.381604 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 15:48:47 crc kubenswrapper[4688]: E1001 15:48:47.381446 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 15:48:47 crc kubenswrapper[4688]: E1001 15:48:47.381709 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5sm44" podUID="c987bdd3-e8bc-473e-99b2-61a143a95543" Oct 01 15:48:47 crc kubenswrapper[4688]: E1001 15:48:47.381805 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.028331 4688 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.028555 4688 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.083334 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.084264 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.085339 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.085980 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.087096 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mbmn8"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.087890 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.088826 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.089847 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.089846 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.100711 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.101494 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.101868 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.101952 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.102505 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.103588 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.104797 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.105788 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.106292 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.107796 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.109414 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.110243 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.110398 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.110857 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.114037 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.127636 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.127837 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.127918 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.128472 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129204 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129273 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zzw65"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129410 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129644 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129851 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.130009 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.130116 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.130337 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-n6cl2"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.129857 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.130736 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.130943 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.131093 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.131178 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.133579 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lkrfg"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.134393 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.137142 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.139103 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.139307 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.140576 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jrw8q"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.141007 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.143053 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tt75q"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.146212 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.146465 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.146498 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.146695 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.146808 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.147018 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.147129 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.147665 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.147770 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.147865 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148008 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148098 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148179 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148255 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148656 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148832 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148867 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148946 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148979 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.148953 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149106 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149380 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149511 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149658 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149836 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.150743 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.149385 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.151163 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.151379 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.152900 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.153324 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-szkjs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.154584 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tlmff"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155008 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155221 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155241 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155363 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155385 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155425 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155384 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155560 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155623 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.155718 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.159684 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.160100 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.160371 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.160603 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.162477 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.162674 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.163318 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.164218 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.166069 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.166794 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.167407 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.167632 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.167987 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.168201 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.168751 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.183930 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc442344-1f3c-4a09-a771-9283daeb07f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.183983 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184018 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-config\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184043 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184064 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5de73e69-cb3e-4d2b-81b3-4ea56231536e-machine-approver-tls\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184085 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184106 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea603264-026d-4e72-aabd-e7d85fe516d8-serving-cert\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184125 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89dz6\" (UniqueName: \"kubernetes.io/projected/ea603264-026d-4e72-aabd-e7d85fe516d8-kube-api-access-89dz6\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184146 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq6qq\" (UniqueName: \"kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184182 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184207 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184225 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184247 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184273 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-auth-proxy-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184294 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184313 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184353 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184375 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfntc\" (UniqueName: \"kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184408 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6jtp\" (UniqueName: \"kubernetes.io/projected/5de73e69-cb3e-4d2b-81b3-4ea56231536e-kube-api-access-j6jtp\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184429 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc442344-1f3c-4a09-a771-9283daeb07f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.184475 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqs78\" (UniqueName: \"kubernetes.io/projected/fc442344-1f3c-4a09-a771-9283daeb07f0-kube-api-access-rqs78\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.188902 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.190763 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.190807 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.191772 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.194072 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.194742 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195000 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195147 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195175 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195259 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195286 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195387 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195430 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195478 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195484 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195563 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195660 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195677 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195788 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195850 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.195902 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.194753 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196097 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196118 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196282 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196595 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196762 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.196785 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.198049 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.199100 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.199204 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209041 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209168 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209307 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209357 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209459 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.209727 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.212582 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.212851 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-n6cl2"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.212867 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mbmn8"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.212877 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.213164 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.213352 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.213485 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.213860 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.214364 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.216638 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.216981 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.219186 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-v4x78"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.219675 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.221014 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.222962 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.223131 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.225141 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.228544 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.229548 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.229586 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.233151 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lx9hp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.233659 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5sgb5"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.233797 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.233799 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.234344 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-992xd"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.234763 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.235041 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.235212 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.235218 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2t92z"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.235519 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.235997 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.236214 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.236552 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.237121 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.237545 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.239549 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.239911 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.240522 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.240943 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.241310 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.241751 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.246357 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-szkjs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.246404 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jrw8q"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.246415 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.263013 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.265058 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.265844 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.266844 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.269379 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.277349 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.277406 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.277416 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zzw65"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.284183 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285124 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-config\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285245 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285362 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-images\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285453 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285561 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqw7d\" (UniqueName: \"kubernetes.io/projected/8ca8a836-be9b-4c80-b8d5-43668ca982a5-kube-api-access-bqw7d\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285667 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b85786e-8bab-4b2c-a908-977c8d0101c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285761 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5de73e69-cb3e-4d2b-81b3-4ea56231536e-machine-approver-tls\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285855 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285957 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286057 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea603264-026d-4e72-aabd-e7d85fe516d8-serving-cert\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286152 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89dz6\" (UniqueName: \"kubernetes.io/projected/ea603264-026d-4e72-aabd-e7d85fe516d8-kube-api-access-89dz6\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286257 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq6qq\" (UniqueName: \"kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286357 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-encryption-config\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286463 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286603 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286717 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-policies\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286832 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286938 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-srv-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287031 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287120 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287217 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-client\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287311 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287416 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287568 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-profile-collector-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287720 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287837 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rhmz\" (UniqueName: \"kubernetes.io/projected/a1f921e5-85a7-4296-bb0d-587c064b48d9-kube-api-access-9rhmz\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.287973 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-config\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288081 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-config\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288181 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-auth-proxy-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288278 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288387 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288486 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285897 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288651 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288774 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288852 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288952 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-serving-cert\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289018 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289113 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-968vx\" (UniqueName: \"kubernetes.io/projected/ce79220d-2d22-4803-afcb-af02069a0f8b-kube-api-access-968vx\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289215 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfntc\" (UniqueName: \"kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289293 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289317 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289439 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-serving-cert\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289574 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6jtp\" (UniqueName: \"kubernetes.io/projected/5de73e69-cb3e-4d2b-81b3-4ea56231536e-kube-api-access-j6jtp\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289678 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.285882 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-config\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.286263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289911 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f53e292-fdef-4716-b217-0c4fab2dd64c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290004 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc442344-1f3c-4a09-a771-9283daeb07f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290126 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzdz\" (UniqueName: \"kubernetes.io/projected/1228ad7f-0f56-4610-ba38-8cb9f64b4e0f-kube-api-access-qxzdz\") pod \"migrator-59844c95c7-npqrk\" (UID: \"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290230 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f921e5-85a7-4296-bb0d-587c064b48d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290328 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290418 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290510 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdbwx\" (UniqueName: \"kubernetes.io/projected/5cd140b8-9734-4f3f-947d-44937d49ffa1-kube-api-access-bdbwx\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290638 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqs78\" (UniqueName: \"kubernetes.io/projected/fc442344-1f3c-4a09-a771-9283daeb07f0-kube-api-access-rqs78\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290733 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290740 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b85786e-8bab-4b2c-a908-977c8d0101c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290808 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-client\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290834 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9lkw\" (UniqueName: \"kubernetes.io/projected/2f53e292-fdef-4716-b217-0c4fab2dd64c-kube-api-access-l9lkw\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290858 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1f921e5-85a7-4296-bb0d-587c064b48d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290891 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc442344-1f3c-4a09-a771-9283daeb07f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290916 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjctn\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-kube-api-access-xjctn\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290946 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdkt2\" (UniqueName: \"kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.290973 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-dir\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.291003 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.291024 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-service-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.289752 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5de73e69-cb3e-4d2b-81b3-4ea56231536e-auth-proxy-config\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.292228 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc442344-1f3c-4a09-a771-9283daeb07f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.288547 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea603264-026d-4e72-aabd-e7d85fe516d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.298011 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.302800 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5de73e69-cb3e-4d2b-81b3-4ea56231536e-machine-approver-tls\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.303779 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.304622 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc442344-1f3c-4a09-a771-9283daeb07f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.304992 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.305316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea603264-026d-4e72-aabd-e7d85fe516d8-serving-cert\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.305775 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.306053 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lkrfg"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.307296 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v4x78"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.315219 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.317396 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.319346 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.322352 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pmzg7"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.323065 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.326043 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.329589 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lx9hp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.330693 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.331694 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.334048 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.335397 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tlmff"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.335577 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.337031 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-992xd"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.337553 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.338890 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.341734 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.346997 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.349329 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.354257 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.355596 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.371780 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tt75q"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.371883 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.371950 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.372015 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2t92z"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.372073 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pmzg7"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.372132 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.372190 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.372289 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-24qdq"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.371386 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.374314 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs55m"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.374444 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.375605 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.375630 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs55m"] Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.375673 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.386965 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.391878 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-config\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.391912 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-config\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.391938 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.391958 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-serving-cert\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.391980 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-968vx\" (UniqueName: \"kubernetes.io/projected/ce79220d-2d22-4803-afcb-af02069a0f8b-kube-api-access-968vx\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392007 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-serving-cert\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392036 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392058 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f53e292-fdef-4716-b217-0c4fab2dd64c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392089 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzdz\" (UniqueName: \"kubernetes.io/projected/1228ad7f-0f56-4610-ba38-8cb9f64b4e0f-kube-api-access-qxzdz\") pod \"migrator-59844c95c7-npqrk\" (UID: \"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392110 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f921e5-85a7-4296-bb0d-587c064b48d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392147 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392165 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392183 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdbwx\" (UniqueName: \"kubernetes.io/projected/5cd140b8-9734-4f3f-947d-44937d49ffa1-kube-api-access-bdbwx\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392288 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b85786e-8bab-4b2c-a908-977c8d0101c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392364 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1f921e5-85a7-4296-bb0d-587c064b48d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392425 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-client\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392631 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9lkw\" (UniqueName: \"kubernetes.io/projected/2f53e292-fdef-4716-b217-0c4fab2dd64c-kube-api-access-l9lkw\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392661 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjctn\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-kube-api-access-xjctn\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392705 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdkt2\" (UniqueName: \"kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392726 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-dir\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392747 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-service-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392768 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392788 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-images\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392865 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqw7d\" (UniqueName: \"kubernetes.io/projected/8ca8a836-be9b-4c80-b8d5-43668ca982a5-kube-api-access-bqw7d\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392889 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b85786e-8bab-4b2c-a908-977c8d0101c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392911 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392945 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-encryption-config\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392966 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.392991 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393012 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-policies\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393029 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-srv-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393062 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-client\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393203 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393229 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-profile-collector-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393256 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.393281 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rhmz\" (UniqueName: \"kubernetes.io/projected/a1f921e5-85a7-4296-bb0d-587c064b48d9-kube-api-access-9rhmz\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.394240 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-config\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.394242 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-dir\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.395200 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.395779 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1f921e5-85a7-4296-bb0d-587c064b48d9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.396229 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.397606 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-config\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.398098 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-serving-cert\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.398877 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-serving-cert\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.399774 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.399791 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-encryption-config\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.400229 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-audit-policies\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.400415 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-service-ca\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.401068 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.401185 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.402204 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.405959 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f53e292-fdef-4716-b217-0c4fab2dd64c-images\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.406146 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.406202 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.407582 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.408499 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f53e292-fdef-4716-b217-0c4fab2dd64c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.409115 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.409854 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce79220d-2d22-4803-afcb-af02069a0f8b-etcd-client\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.410050 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1f921e5-85a7-4296-bb0d-587c064b48d9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.417451 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ca8a836-be9b-4c80-b8d5-43668ca982a5-etcd-client\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.426084 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.446542 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.466118 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.485937 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.506683 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.520048 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b85786e-8bab-4b2c-a908-977c8d0101c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.539780 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.544414 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b85786e-8bab-4b2c-a908-977c8d0101c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.547671 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.566778 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.587979 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.606747 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.630714 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.673805 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.687510 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.706648 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.726324 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.746300 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.767518 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.786449 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.807575 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.828242 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.847211 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.867676 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.887158 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.908004 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.927461 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.948146 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.966706 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 01 15:48:48 crc kubenswrapper[4688]: I1001 15:48:48.988764 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.007456 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.027905 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.048612 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.067621 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.086901 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.094446 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-srv-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.106276 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.115249 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd140b8-9734-4f3f-947d-44937d49ffa1-profile-collector-cert\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.126941 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.146765 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.166977 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.187552 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.227868 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.245708 4688 request.go:700] Waited for 1.011514617s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-ac-dockercfg-9lkdf&limit=500&resourceVersion=0 Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.248023 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.267384 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.287732 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.307068 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.326658 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.347129 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.366425 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.380609 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.380670 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.380681 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.381024 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.387385 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.406713 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.427456 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.446133 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.467059 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.486348 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.506459 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.527004 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.547200 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.567284 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.587068 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.607592 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.627623 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.647888 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.667108 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.686471 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.707281 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.727355 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.756431 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.766940 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.787723 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.806969 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.828142 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.846859 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.866953 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.887739 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.907818 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.927652 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.978998 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq6qq\" (UniqueName: \"kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq\") pod \"controller-manager-879f6c89f-xrtrx\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.991963 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:49 crc kubenswrapper[4688]: I1001 15:48:49.997273 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89dz6\" (UniqueName: \"kubernetes.io/projected/ea603264-026d-4e72-aabd-e7d85fe516d8-kube-api-access-89dz6\") pod \"authentication-operator-69f744f599-mbmn8\" (UID: \"ea603264-026d-4e72-aabd-e7d85fe516d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.025177 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfntc\" (UniqueName: \"kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc\") pod \"route-controller-manager-6576b87f9c-gb5hh\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.026303 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6jtp\" (UniqueName: \"kubernetes.io/projected/5de73e69-cb3e-4d2b-81b3-4ea56231536e-kube-api-access-j6jtp\") pod \"machine-approver-56656f9798-dkwrm\" (UID: \"5de73e69-cb3e-4d2b-81b3-4ea56231536e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.048146 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.053298 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqs78\" (UniqueName: \"kubernetes.io/projected/fc442344-1f3c-4a09-a771-9283daeb07f0-kube-api-access-rqs78\") pod \"openshift-controller-manager-operator-756b6f6bc6-2tdnx\" (UID: \"fc442344-1f3c-4a09-a771-9283daeb07f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.067071 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.087760 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.107133 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.128792 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.147665 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.169243 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.187744 4688 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.206666 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.228072 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.234810 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.242022 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.246598 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:50 crc kubenswrapper[4688]: W1001 15:48:50.263838 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5de73e69_cb3e_4d2b_81b3_4ea56231536e.slice/crio-bc10b599fde1e8d9dc805bb684d51023ae1eb9d46ae877af7d5626ecc54ddebc WatchSource:0}: Error finding container bc10b599fde1e8d9dc805bb684d51023ae1eb9d46ae877af7d5626ecc54ddebc: Status 404 returned error can't find the container with id bc10b599fde1e8d9dc805bb684d51023ae1eb9d46ae877af7d5626ecc54ddebc Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.264835 4688 request.go:700] Waited for 1.871417111s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.268117 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.274354 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdbwx\" (UniqueName: \"kubernetes.io/projected/5cd140b8-9734-4f3f-947d-44937d49ffa1-kube-api-access-bdbwx\") pod \"catalog-operator-68c6474976-f8thq\" (UID: \"5cd140b8-9734-4f3f-947d-44937d49ffa1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.290756 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rhmz\" (UniqueName: \"kubernetes.io/projected/a1f921e5-85a7-4296-bb0d-587c064b48d9-kube-api-access-9rhmz\") pod \"openshift-apiserver-operator-796bbdcf4f-kdnf8\" (UID: \"a1f921e5-85a7-4296-bb0d-587c064b48d9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.293097 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.314231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9lkw\" (UniqueName: \"kubernetes.io/projected/2f53e292-fdef-4716-b217-0c4fab2dd64c-kube-api-access-l9lkw\") pod \"machine-api-operator-5694c8668f-szkjs\" (UID: \"2f53e292-fdef-4716-b217-0c4fab2dd64c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.317852 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.348876 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjctn\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-kube-api-access-xjctn\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.393835 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdkt2\" (UniqueName: \"kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2\") pod \"console-f9d7485db-8r5l7\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.405598 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzdz\" (UniqueName: \"kubernetes.io/projected/1228ad7f-0f56-4610-ba38-8cb9f64b4e0f-kube-api-access-qxzdz\") pod \"migrator-59844c95c7-npqrk\" (UID: \"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.406039 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b85786e-8bab-4b2c-a908-977c8d0101c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5zwvb\" (UID: \"7b85786e-8bab-4b2c-a908-977c8d0101c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.408586 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-968vx\" (UniqueName: \"kubernetes.io/projected/ce79220d-2d22-4803-afcb-af02069a0f8b-kube-api-access-968vx\") pod \"etcd-operator-b45778765-jrw8q\" (UID: \"ce79220d-2d22-4803-afcb-af02069a0f8b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.425596 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqw7d\" (UniqueName: \"kubernetes.io/projected/8ca8a836-be9b-4c80-b8d5-43668ca982a5-kube-api-access-bqw7d\") pod \"apiserver-7bbb656c7d-zgxfs\" (UID: \"8ca8a836-be9b-4c80-b8d5-43668ca982a5\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.440776 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.465553 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.467297 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.488315 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.498846 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.499967 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.510268 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.526494 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.527646 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532034 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532064 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpntl\" (UniqueName: \"kubernetes.io/projected/7b007c86-dbed-47a0-b456-aa6be1e6b21c-kube-api-access-fpntl\") pod \"downloads-7954f5f757-n6cl2\" (UID: \"7b007c86-dbed-47a0-b456-aa6be1e6b21c\") " pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532157 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f98bf7-aac6-4e31-8ebd-9237db04f612-config\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532188 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79546\" (UniqueName: \"kubernetes.io/projected/042bc31a-ee11-4740-baa9-52474d0e273f-kube-api-access-79546\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532240 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmqgz\" (UniqueName: \"kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.532289 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2814f7d1-706b-450a-b47c-2ccc77ee6c20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533226 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb1e661-4852-4056-950f-90a2951e762e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533255 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2814f7d1-706b-450a-b47c-2ccc77ee6c20-config\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533279 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533303 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-metrics-tls\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533352 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmbx7\" (UniqueName: \"kubernetes.io/projected/2b776271-b6a4-42a2-b352-8b474016fb30-kube-api-access-kmbx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533386 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qt9r\" (UniqueName: \"kubernetes.io/projected/41a72e46-e9a8-48f0-b002-b8930f2283a9-kube-api-access-6qt9r\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533410 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-audit\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533436 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca5f8204-18df-4b10-ab58-0ea02df49e4f-serving-cert\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533472 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533496 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlrxr\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533515 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2814f7d1-706b-450a-b47c-2ccc77ee6c20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533868 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533907 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-node-pullsecrets\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533936 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533970 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb1e661-4852-4056-950f-90a2951e762e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.533992 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abb1e661-4852-4056-950f-90a2951e762e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534015 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/41a72e46-e9a8-48f0-b002-b8930f2283a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534043 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrptp\" (UniqueName: \"kubernetes.io/projected/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-kube-api-access-vrptp\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534064 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534376 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.534502 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.034492001 +0000 UTC m=+120.385131963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534499 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534557 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534850 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-image-import-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534880 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd565\" (UniqueName: \"kubernetes.io/projected/4b66d47b-5d72-45ad-bb20-675829bf567a-kube-api-access-bd565\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.534993 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/042bc31a-ee11-4740-baa9-52474d0e273f-metrics-tls\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535016 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-serving-cert\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535038 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535058 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-encryption-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535101 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17f98bf7-aac6-4e31-8ebd-9237db04f612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535149 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535176 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535191 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535204 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535220 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-config\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535234 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-client\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535352 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b776271-b6a4-42a2-b352-8b474016fb30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535370 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-trusted-ca\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535388 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a426294-bb9a-4532-a89e-36464ead5a80-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535403 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535470 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535513 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f98bf7-aac6-4e31-8ebd-9237db04f612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535638 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535657 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535677 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-serving-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535695 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535732 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535775 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b776271-b6a4-42a2-b352-8b474016fb30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535798 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535818 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535839 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgt59\" (UniqueName: \"kubernetes.io/projected/ca5f8204-18df-4b10-ab58-0ea02df49e4f-kube-api-access-tgt59\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535858 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e16847f2-730e-4ca5-a193-fde4db19cb2c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535877 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535896 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh49f\" (UniqueName: \"kubernetes.io/projected/e16847f2-730e-4ca5-a193-fde4db19cb2c-kube-api-access-lh49f\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535918 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgrzk\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-kube-api-access-wgrzk\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535953 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-audit-dir\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535971 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a72e46-e9a8-48f0-b002-b8930f2283a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.535990 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a426294-bb9a-4532-a89e-36464ead5a80-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.536011 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-config-volume\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.544932 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.546008 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.572975 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.637914 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.638067 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.138041866 +0000 UTC m=+120.488681828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638127 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79546\" (UniqueName: \"kubernetes.io/projected/042bc31a-ee11-4740-baa9-52474d0e273f-kube-api-access-79546\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638162 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638203 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmqgz\" (UniqueName: \"kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638231 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5flrn\" (UniqueName: \"kubernetes.io/projected/9b6bc4f2-3197-45e8-b934-c666e37c167e-kube-api-access-5flrn\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638301 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2814f7d1-706b-450a-b47c-2ccc77ee6c20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638328 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dabdf9c3-fc17-46fd-bf40-5a67705823cc-cert\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638353 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-csi-data-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638377 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb1e661-4852-4056-950f-90a2951e762e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638401 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2814f7d1-706b-450a-b47c-2ccc77ee6c20-config\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638424 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25ndb\" (UniqueName: \"kubernetes.io/projected/47fa404e-12b7-4590-8a04-7b304962b3ea-kube-api-access-25ndb\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638446 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3f387ea-c091-4def-a25f-f299bb8ee3f9-proxy-tls\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638469 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-metrics-tls\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638496 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638557 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638811 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmbx7\" (UniqueName: \"kubernetes.io/projected/2b776271-b6a4-42a2-b352-8b474016fb30-kube-api-access-kmbx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638843 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qt9r\" (UniqueName: \"kubernetes.io/projected/41a72e46-e9a8-48f0-b002-b8930f2283a9-kube-api-access-6qt9r\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.638880 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-service-ca-bundle\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.639302 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2814f7d1-706b-450a-b47c-2ccc77ee6c20-config\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.639400 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-audit\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.639665 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca5f8204-18df-4b10-ab58-0ea02df49e4f-serving-cert\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.639701 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5433040b-8666-4f49-98ce-9129ef301164-signing-key\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640697 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640739 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlrxr\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640762 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3d403b13-9b49-4193-9589-4f29e68ea9a6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640779 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glkqp\" (UniqueName: \"kubernetes.io/projected/4a8ec615-416d-4e20-b133-3adfc59692db-kube-api-access-glkqp\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640797 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640815 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-node-bootstrap-token\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640829 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-metrics-certs\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640870 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640885 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-default-certificate\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640901 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-config\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640922 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkbh8\" (UniqueName: \"kubernetes.io/projected/3fa56b51-5928-435f-887a-13adad0cf06e-kube-api-access-fkbh8\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640944 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2814f7d1-706b-450a-b47c-2ccc77ee6c20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640979 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.640996 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-node-pullsecrets\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641030 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgbwv\" (UniqueName: \"kubernetes.io/projected/dc8ba611-f5be-48f7-990f-f56d4f7a9539-kube-api-access-zgbwv\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641054 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641072 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641095 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb1e661-4852-4056-950f-90a2951e762e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641110 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abb1e661-4852-4056-950f-90a2951e762e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641124 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/41a72e46-e9a8-48f0-b002-b8930f2283a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641140 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6bc4f2-3197-45e8-b934-c666e37c167e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641157 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5kmq\" (UniqueName: \"kubernetes.io/projected/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-kube-api-access-c5kmq\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641173 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641190 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13ad828f-f06e-48a3-a2ba-bb6f757b356e-tmpfs\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641205 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrptp\" (UniqueName: \"kubernetes.io/projected/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-kube-api-access-vrptp\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641230 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641245 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641270 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-image-import-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641289 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd565\" (UniqueName: \"kubernetes.io/projected/4b66d47b-5d72-45ad-bb20-675829bf567a-kube-api-access-bd565\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641311 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6xv8\" (UniqueName: \"kubernetes.io/projected/d3f387ea-c091-4def-a25f-f299bb8ee3f9-kube-api-access-l6xv8\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641335 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfgtb\" (UniqueName: \"kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641351 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-stats-auth\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641390 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc8ba611-f5be-48f7-990f-f56d4f7a9539-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641407 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-socket-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641434 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/042bc31a-ee11-4740-baa9-52474d0e273f-metrics-tls\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641455 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-serving-cert\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.641496 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.642562 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.642593 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.642781 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-audit\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.643492 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-encryption-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.644401 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.644597 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.644795 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.647692 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca5f8204-18df-4b10-ab58-0ea02df49e4f-serving-cert\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.648413 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abb1e661-4852-4056-950f-90a2951e762e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkr4m\" (UniqueName: \"kubernetes.io/projected/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-kube-api-access-xkr4m\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649116 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-images\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649161 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17f98bf7-aac6-4e31-8ebd-9237db04f612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649233 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-webhook-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649260 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649312 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.649336 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-plugins-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.650827 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-image-import-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.653660 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.153642333 +0000 UTC m=+120.504282295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.653960 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-node-pullsecrets\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659603 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659637 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659665 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659685 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-client\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659707 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-config\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659783 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-registration-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659815 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b776271-b6a4-42a2-b352-8b474016fb30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659833 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-trusted-ca\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659851 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a426294-bb9a-4532-a89e-36464ead5a80-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659868 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659898 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-mountpoint-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659929 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.659945 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.662861 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.663249 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abb1e661-4852-4056-950f-90a2951e762e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.663275 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.663299 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.666816 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-metrics-tls\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.667335 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-config\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.667640 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca5f8204-18df-4b10-ab58-0ea02df49e4f-trusted-ca\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.667834 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.667906 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.669012 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f98bf7-aac6-4e31-8ebd-9237db04f612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.669044 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.669072 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.674647 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.674720 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-serving-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.674748 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.674818 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgt9c\" (UniqueName: \"kubernetes.io/projected/dabdf9c3-fc17-46fd-bf40-5a67705823cc-kube-api-access-wgt9c\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.675363 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.677798 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/41a72e46-e9a8-48f0-b002-b8930f2283a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.678261 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-serving-ca\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.678494 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-encryption-config\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679366 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679368 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679430 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sd8h\" (UniqueName: \"kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-srv-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679714 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679767 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b776271-b6a4-42a2-b352-8b474016fb30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679794 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.679837 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681204 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681225 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlkl8\" (UniqueName: \"kubernetes.io/projected/13ad828f-f06e-48a3-a2ba-bb6f757b356e-kube-api-access-jlkl8\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681239 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-certs\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681258 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqrps\" (UniqueName: \"kubernetes.io/projected/3d403b13-9b49-4193-9589-4f29e68ea9a6-kube-api-access-qqrps\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681275 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-serving-cert\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681295 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgt59\" (UniqueName: \"kubernetes.io/projected/ca5f8204-18df-4b10-ab58-0ea02df49e4f-kube-api-access-tgt59\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681310 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e16847f2-730e-4ca5-a193-fde4db19cb2c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681327 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681342 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh49f\" (UniqueName: \"kubernetes.io/projected/e16847f2-730e-4ca5-a193-fde4db19cb2c-kube-api-access-lh49f\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681359 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-proxy-tls\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681396 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgrzk\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-kube-api-access-wgrzk\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681427 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-audit-dir\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681441 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a72e46-e9a8-48f0-b002-b8930f2283a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681457 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a426294-bb9a-4532-a89e-36464ead5a80-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681472 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgvph\" (UniqueName: \"kubernetes.io/projected/5433040b-8666-4f49-98ce-9129ef301164-kube-api-access-sgvph\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681501 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-config-volume\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681523 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpntl\" (UniqueName: \"kubernetes.io/projected/7b007c86-dbed-47a0-b456-aa6be1e6b21c-kube-api-access-fpntl\") pod \"downloads-7954f5f757-n6cl2\" (UID: \"7b007c86-dbed-47a0-b456-aa6be1e6b21c\") " pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681566 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f98bf7-aac6-4e31-8ebd-9237db04f612-config\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681584 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxncn\" (UniqueName: \"kubernetes.io/projected/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-kube-api-access-hxncn\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.681606 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5433040b-8666-4f49-98ce-9129ef301164-signing-cabundle\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.682496 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.683193 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b66d47b-5d72-45ad-bb20-675829bf567a-audit-dir\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.683475 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-config-volume\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.684370 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f98bf7-aac6-4e31-8ebd-9237db04f612-config\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.684432 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a426294-bb9a-4532-a89e-36464ead5a80-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.684861 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-serving-cert\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.686004 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mbmn8"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.686179 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b776271-b6a4-42a2-b352-8b474016fb30-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.686450 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.687947 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.690465 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e16847f2-730e-4ca5-a193-fde4db19cb2c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.690467 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/042bc31a-ee11-4740-baa9-52474d0e273f-metrics-tls\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.690996 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.691176 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41a72e46-e9a8-48f0-b002-b8930f2283a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.691602 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.692840 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17f98bf7-aac6-4e31-8ebd-9237db04f612-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.695623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a426294-bb9a-4532-a89e-36464ead5a80-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.699291 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b66d47b-5d72-45ad-bb20-675829bf567a-etcd-client\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.700042 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.706124 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.706878 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmqgz\" (UniqueName: \"kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.707488 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.708763 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.708998 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m94rs\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.709756 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2814f7d1-706b-450a-b47c-2ccc77ee6c20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.710782 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b776271-b6a4-42a2-b352-8b474016fb30-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.713970 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79546\" (UniqueName: \"kubernetes.io/projected/042bc31a-ee11-4740-baa9-52474d0e273f-kube-api-access-79546\") pod \"dns-operator-744455d44c-zzw65\" (UID: \"042bc31a-ee11-4740-baa9-52474d0e273f\") " pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.738718 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2814f7d1-706b-450a-b47c-2ccc77ee6c20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q2vld\" (UID: \"2814f7d1-706b-450a-b47c-2ccc77ee6c20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.743972 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qt9r\" (UniqueName: \"kubernetes.io/projected/41a72e46-e9a8-48f0-b002-b8930f2283a9-kube-api-access-6qt9r\") pod \"openshift-config-operator-7777fb866f-tlmff\" (UID: \"41a72e46-e9a8-48f0-b002-b8930f2283a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.750041 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.773349 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmbx7\" (UniqueName: \"kubernetes.io/projected/2b776271-b6a4-42a2-b352-8b474016fb30-kube-api-access-kmbx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-rv2jj\" (UID: \"2b776271-b6a4-42a2-b352-8b474016fb30\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.782198 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.786011 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrptp\" (UniqueName: \"kubernetes.io/projected/c8d1bcc4-727d-4e91-a6c6-3658cee8baa1-kube-api-access-vrptp\") pod \"dns-default-v4x78\" (UID: \"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1\") " pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787442 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787624 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkr4m\" (UniqueName: \"kubernetes.io/projected/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-kube-api-access-xkr4m\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787650 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-images\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787694 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-webhook-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787722 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-plugins-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787744 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787767 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-registration-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787789 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-mountpoint-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787812 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787850 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgt9c\" (UniqueName: \"kubernetes.io/projected/dabdf9c3-fc17-46fd-bf40-5a67705823cc-kube-api-access-wgt9c\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787872 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787892 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sd8h\" (UniqueName: \"kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787912 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-srv-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787949 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787973 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlkl8\" (UniqueName: \"kubernetes.io/projected/13ad828f-f06e-48a3-a2ba-bb6f757b356e-kube-api-access-jlkl8\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.787992 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-certs\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788013 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqrps\" (UniqueName: \"kubernetes.io/projected/3d403b13-9b49-4193-9589-4f29e68ea9a6-kube-api-access-qqrps\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788033 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-serving-cert\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788066 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-proxy-tls\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788104 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgvph\" (UniqueName: \"kubernetes.io/projected/5433040b-8666-4f49-98ce-9129ef301164-kube-api-access-sgvph\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788140 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxncn\" (UniqueName: \"kubernetes.io/projected/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-kube-api-access-hxncn\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788163 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5433040b-8666-4f49-98ce-9129ef301164-signing-cabundle\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788187 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788213 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5flrn\" (UniqueName: \"kubernetes.io/projected/9b6bc4f2-3197-45e8-b934-c666e37c167e-kube-api-access-5flrn\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788235 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dabdf9c3-fc17-46fd-bf40-5a67705823cc-cert\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788254 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-csi-data-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788273 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25ndb\" (UniqueName: \"kubernetes.io/projected/47fa404e-12b7-4590-8a04-7b304962b3ea-kube-api-access-25ndb\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788293 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3f387ea-c091-4def-a25f-f299bb8ee3f9-proxy-tls\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788314 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788340 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-service-ca-bundle\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788351 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-registration-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5433040b-8666-4f49-98ce-9129ef301164-signing-key\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788395 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-mountpoint-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788395 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3d403b13-9b49-4193-9589-4f29e68ea9a6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788420 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glkqp\" (UniqueName: \"kubernetes.io/projected/4a8ec615-416d-4e20-b133-3adfc59692db-kube-api-access-glkqp\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788438 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-node-bootstrap-token\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788452 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-metrics-certs\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788475 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788491 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-default-certificate\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.788506 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-config\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.793133 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.793230 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.293212254 +0000 UTC m=+120.643852216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.793793 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-images\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.794328 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-plugins-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.794659 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3d403b13-9b49-4193-9589-4f29e68ea9a6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.795475 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797282 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkbh8\" (UniqueName: \"kubernetes.io/projected/3fa56b51-5928-435f-887a-13adad0cf06e-kube-api-access-fkbh8\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797337 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797360 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgbwv\" (UniqueName: \"kubernetes.io/projected/dc8ba611-f5be-48f7-990f-f56d4f7a9539-kube-api-access-zgbwv\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797378 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797404 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6bc4f2-3197-45e8-b934-c666e37c167e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797421 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5kmq\" (UniqueName: \"kubernetes.io/projected/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-kube-api-access-c5kmq\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797441 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13ad828f-f06e-48a3-a2ba-bb6f757b356e-tmpfs\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797482 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6xv8\" (UniqueName: \"kubernetes.io/projected/d3f387ea-c091-4def-a25f-f299bb8ee3f9-kube-api-access-l6xv8\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797501 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfgtb\" (UniqueName: \"kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797521 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-stats-auth\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797591 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc8ba611-f5be-48f7-990f-f56d4f7a9539-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797608 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-socket-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.797729 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-socket-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.798172 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-config\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.798472 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.298461438 +0000 UTC m=+120.649101400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.799316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.800245 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5433040b-8666-4f49-98ce-9129ef301164-signing-cabundle\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.802227 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3f387ea-c091-4def-a25f-f299bb8ee3f9-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.803271 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/13ad828f-f06e-48a3-a2ba-bb6f757b356e-tmpfs\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.804191 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8ec615-416d-4e20-b133-3adfc59692db-csi-data-dir\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.808860 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-service-ca-bundle\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.824369 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-webhook-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.825048 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.827921 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-srv-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.831397 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-stats-auth\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.831822 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-node-bootstrap-token\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.832470 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/13ad828f-f06e-48a3-a2ba-bb6f757b356e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.833551 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.834009 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47fa404e-12b7-4590-8a04-7b304962b3ea-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.836317 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-default-certificate\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.839980 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-metrics-certs\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.843803 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.861438 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-serving-cert\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.861574 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd565\" (UniqueName: \"kubernetes.io/projected/4b66d47b-5d72-45ad-bb20-675829bf567a-kube-api-access-bd565\") pod \"apiserver-76f77b778f-lkrfg\" (UID: \"4b66d47b-5d72-45ad-bb20-675829bf567a\") " pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.861796 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b6bc4f2-3197-45e8-b934-c666e37c167e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.862635 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5433040b-8666-4f49-98ce-9129ef301164-signing-key\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.863898 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dabdf9c3-fc17-46fd-bf40-5a67705823cc-cert\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.864209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3f387ea-c091-4def-a25f-f299bb8ee3f9-proxy-tls\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.866288 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abb1e661-4852-4056-950f-90a2951e762e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nkp4l\" (UID: \"abb1e661-4852-4056-950f-90a2951e762e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.871823 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.871863 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-proxy-tls\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.873496 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3fa56b51-5928-435f-887a-13adad0cf06e-certs\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.874286 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc8ba611-f5be-48f7-990f-f56d4f7a9539-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.876033 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlrxr\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.879797 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.887213 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.891504 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f98bf7-aac6-4e31-8ebd-9237db04f612-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gmsfp\" (UID: \"17f98bf7-aac6-4e31-8ebd-9237db04f612\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.901964 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:50 crc kubenswrapper[4688]: E1001 15:48:50.902928 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.40291376 +0000 UTC m=+120.753553722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.907649 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.909003 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.936313 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgt59\" (UniqueName: \"kubernetes.io/projected/ca5f8204-18df-4b10-ab58-0ea02df49e4f-kube-api-access-tgt59\") pod \"console-operator-58897d9998-tt75q\" (UID: \"ca5f8204-18df-4b10-ab58-0ea02df49e4f\") " pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.944899 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgrzk\" (UniqueName: \"kubernetes.io/projected/4a426294-bb9a-4532-a89e-36464ead5a80-kube-api-access-wgrzk\") pod \"cluster-image-registry-operator-dc59b4c8b-mhll6\" (UID: \"4a426294-bb9a-4532-a89e-36464ead5a80\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.964050 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpntl\" (UniqueName: \"kubernetes.io/projected/7b007c86-dbed-47a0-b456-aa6be1e6b21c-kube-api-access-fpntl\") pod \"downloads-7954f5f757-n6cl2\" (UID: \"7b007c86-dbed-47a0-b456-aa6be1e6b21c\") " pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.974741 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb"] Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.984681 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh49f\" (UniqueName: \"kubernetes.io/projected/e16847f2-730e-4ca5-a193-fde4db19cb2c-kube-api-access-lh49f\") pod \"cluster-samples-operator-665b6dd947-nd6cx\" (UID: \"e16847f2-730e-4ca5-a193-fde4db19cb2c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:50 crc kubenswrapper[4688]: I1001 15:48:50.985472 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.004279 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.004753 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.504741424 +0000 UTC m=+120.855381386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.005959 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.026463 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.029653 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.030352 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sd8h\" (UniqueName: \"kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h\") pod \"marketplace-operator-79b997595-vfflp\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.050352 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkr4m\" (UniqueName: \"kubernetes.io/projected/90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea-kube-api-access-xkr4m\") pod \"machine-config-controller-84d6567774-992xd\" (UID: \"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.062544 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.065630 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jrw8q"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.106455 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.106666 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgt9c\" (UniqueName: \"kubernetes.io/projected/dabdf9c3-fc17-46fd-bf40-5a67705823cc-kube-api-access-wgt9c\") pod \"ingress-canary-pmzg7\" (UID: \"dabdf9c3-fc17-46fd-bf40-5a67705823cc\") " pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.106751 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.606737074 +0000 UTC m=+120.957377036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.106894 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.107166 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.607155466 +0000 UTC m=+120.957795418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.108473 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.111639 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.121649 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.123583 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkbh8\" (UniqueName: \"kubernetes.io/projected/3fa56b51-5928-435f-887a-13adad0cf06e-kube-api-access-fkbh8\") pod \"machine-config-server-24qdq\" (UID: \"3fa56b51-5928-435f-887a-13adad0cf06e\") " pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.143919 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.144895 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glkqp\" (UniqueName: \"kubernetes.io/projected/4a8ec615-416d-4e20-b133-3adfc59692db-kube-api-access-glkqp\") pod \"csi-hostpathplugin-zs55m\" (UID: \"4a8ec615-416d-4e20-b133-3adfc59692db\") " pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.151298 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.151595 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgbwv\" (UniqueName: \"kubernetes.io/projected/dc8ba611-f5be-48f7-990f-f56d4f7a9539-kube-api-access-zgbwv\") pod \"control-plane-machine-set-operator-78cbb6b69f-9kj7x\" (UID: \"dc8ba611-f5be-48f7-990f-f56d4f7a9539\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.159067 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.161481 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.167911 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6xv8\" (UniqueName: \"kubernetes.io/projected/d3f387ea-c091-4def-a25f-f299bb8ee3f9-kube-api-access-l6xv8\") pod \"machine-config-operator-74547568cd-sq75x\" (UID: \"d3f387ea-c091-4def-a25f-f299bb8ee3f9\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.169726 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-szkjs"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.172935 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfgtb\" (UniqueName: \"kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb\") pod \"collect-profiles-29322225-nv5w9\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.177574 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs"] Oct 01 15:48:51 crc kubenswrapper[4688]: W1001 15:48:51.179738 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1228ad7f_0f56_4610_ba38_8cb9f64b4e0f.slice/crio-6bb6a055e059b77302affc87baaa00fcae3a83d05899c1153623669411e109e5 WatchSource:0}: Error finding container 6bb6a055e059b77302affc87baaa00fcae3a83d05899c1153623669411e109e5: Status 404 returned error can't find the container with id 6bb6a055e059b77302affc87baaa00fcae3a83d05899c1153623669411e109e5 Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.183498 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5kmq\" (UniqueName: \"kubernetes.io/projected/8699a8ba-1dc6-49df-ae9b-d1feb4db64ec-kube-api-access-c5kmq\") pod \"service-ca-operator-777779d784-lhmpv\" (UID: \"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.186464 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" event={"ID":"5de73e69-cb3e-4d2b-81b3-4ea56231536e","Type":"ContainerStarted","Data":"cc2f403e1e224c5748e0949b1ebb6c71d1d6cffa537ba31bd34b52ec05c4af88"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.186509 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" event={"ID":"5de73e69-cb3e-4d2b-81b3-4ea56231536e","Type":"ContainerStarted","Data":"bc10b599fde1e8d9dc805bb684d51023ae1eb9d46ae877af7d5626ecc54ddebc"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.203886 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25ndb\" (UniqueName: \"kubernetes.io/projected/47fa404e-12b7-4590-8a04-7b304962b3ea-kube-api-access-25ndb\") pod \"olm-operator-6b444d44fb-wzscd\" (UID: \"47fa404e-12b7-4590-8a04-7b304962b3ea\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.207745 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.208182 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.708166217 +0000 UTC m=+121.058806179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.213228 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" event={"ID":"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f","Type":"ContainerStarted","Data":"fa9d92515bb4ba2e1293a8c0e8b5fee660e8d70635dfd2e08cdc41ff14e74138"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.213270 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" event={"ID":"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f","Type":"ContainerStarted","Data":"d25027240e9b41d09ad25a43fafd39af01ac6cb708890d55fe2502e677e175a9"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.214086 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.216701 4688 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xrtrx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.216751 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.220321 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tlmff"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.238891 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" event={"ID":"fa9c4e7c-802a-4983-9b9c-607673fb8e94","Type":"ContainerStarted","Data":"ff72140b58e2bd7ec8cdbcb0e57a1a2a086fc6a448b88cd3c68dff1769304a8c"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.238936 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" event={"ID":"fa9c4e7c-802a-4983-9b9c-607673fb8e94","Type":"ContainerStarted","Data":"fa0630f6ce9db7f0bdd58c04658a5360e032e67ddfd99937b73433d4078663a2"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.239647 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.239978 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.240867 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5flrn\" (UniqueName: \"kubernetes.io/projected/9b6bc4f2-3197-45e8-b934-c666e37c167e-kube-api-access-5flrn\") pod \"package-server-manager-789f6589d5-vxjr5\" (UID: \"9b6bc4f2-3197-45e8-b934-c666e37c167e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.243485 4688 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-gb5hh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.243538 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.246765 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgvph\" (UniqueName: \"kubernetes.io/projected/5433040b-8666-4f49-98ce-9129ef301164-kube-api-access-sgvph\") pod \"service-ca-9c57cc56f-2t92z\" (UID: \"5433040b-8666-4f49-98ce-9129ef301164\") " pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.248730 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" event={"ID":"5cd140b8-9734-4f3f-947d-44937d49ffa1","Type":"ContainerStarted","Data":"d16eed58ec1b0e3ed04a50af2a1b0a3ef7481163fb81f7a015f1977480b63db0"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.252745 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.259389 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.262632 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" event={"ID":"7b85786e-8bab-4b2c-a908-977c8d0101c7","Type":"ContainerStarted","Data":"19145576653ac552e64fc17d1614d5ba623a896cfef04034ba00208ba60758f1"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.267497 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.272072 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.273161 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" event={"ID":"fc442344-1f3c-4a09-a771-9283daeb07f0","Type":"ContainerStarted","Data":"e74cf84e729f07dadf743e8f3fb06d0b775703b075c1c4a534a90ce5cb43b6ad"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.275517 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" event={"ID":"ea603264-026d-4e72-aabd-e7d85fe516d8","Type":"ContainerStarted","Data":"51e708505e2f17323714a7d5583243ceb0b94e66c8187b1e79865fd8156002de"} Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.280229 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxncn\" (UniqueName: \"kubernetes.io/projected/ba2124c7-a29e-4a00-be83-9679fdbe8c7d-kube-api-access-hxncn\") pod \"router-default-5444994796-5sgb5\" (UID: \"ba2124c7-a29e-4a00-be83-9679fdbe8c7d\") " pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.285093 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlkl8\" (UniqueName: \"kubernetes.io/projected/13ad828f-f06e-48a3-a2ba-bb6f757b356e-kube-api-access-jlkl8\") pod \"packageserver-d55dfcdfc-wxhpj\" (UID: \"13ad828f-f06e-48a3-a2ba-bb6f757b356e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.287854 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.297798 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.302555 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqrps\" (UniqueName: \"kubernetes.io/projected/3d403b13-9b49-4193-9589-4f29e68ea9a6-kube-api-access-qqrps\") pod \"multus-admission-controller-857f4d67dd-lx9hp\" (UID: \"3d403b13-9b49-4193-9589-4f29e68ea9a6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.355109 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.365800 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.865785216 +0000 UTC m=+121.216425178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.375759 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.383632 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-24qdq" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.393690 4688 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.393749 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.407221 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.413480 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.426800 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.433742 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pmzg7" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.459934 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.460183 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.960161353 +0000 UTC m=+121.310801315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.460521 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.460915 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:51.960900734 +0000 UTC m=+121.311540696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.466285 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.473913 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.513816 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.522387 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.548150 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.549401 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.559356 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lkrfg"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.561972 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.562153 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.0621288 +0000 UTC m=+121.412768762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.562473 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.562809 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.06279675 +0000 UTC m=+121.413436712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.579176 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.587374 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.588033 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.618812 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.668654 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.668788 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.168772506 +0000 UTC m=+121.519412468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.669141 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.669606 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.16959384 +0000 UTC m=+121.520233802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.691412 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zzw65"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.716513 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.717667 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v4x78"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.772845 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.773157 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.273134365 +0000 UTC m=+121.623774327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.773334 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.773680 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.273669711 +0000 UTC m=+121.624309673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.818574 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.859099 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.873756 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.874661 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.37462853 +0000 UTC m=+121.725268552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.915626 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tt75q"] Oct 01 15:48:51 crc kubenswrapper[4688]: I1001 15:48:51.975069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:51 crc kubenswrapper[4688]: E1001 15:48:51.975380 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.475367252 +0000 UTC m=+121.826007214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.040450 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-n6cl2"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.057841 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.078492 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.079326 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.579309029 +0000 UTC m=+121.929948991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.180077 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" podStartSLOduration=99.180059172 podStartE2EDuration="1m39.180059172s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:52.138153674 +0000 UTC m=+121.488793636" watchObservedRunningTime="2025-10-01 15:48:52.180059172 +0000 UTC m=+121.530699134" Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.180227 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.180486 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.680473654 +0000 UTC m=+122.031113616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: W1001 15:48:52.190289 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b007c86_dbed_47a0_b456_aa6be1e6b21c.slice/crio-875ea16e858dc23e85a51c08b80310ff50a2bf557f20ddb7bdbf88b4818b42c2 WatchSource:0}: Error finding container 875ea16e858dc23e85a51c08b80310ff50a2bf557f20ddb7bdbf88b4818b42c2: Status 404 returned error can't find the container with id 875ea16e858dc23e85a51c08b80310ff50a2bf557f20ddb7bdbf88b4818b42c2 Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.233714 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.281412 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.281837 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.781822205 +0000 UTC m=+122.132462167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.312217 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" event={"ID":"fc442344-1f3c-4a09-a771-9283daeb07f0","Type":"ContainerStarted","Data":"e943bfd23fb41cd42123897b305216089734938499f75a3457fb4d1401d049b8"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.313608 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" event={"ID":"47fa404e-12b7-4590-8a04-7b304962b3ea","Type":"ContainerStarted","Data":"ba8827bbb0f49f49acbe530d19b7b12d3bf7576a4713600026e7b6b48d0d1e26"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.314606 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" event={"ID":"ce79220d-2d22-4803-afcb-af02069a0f8b","Type":"ContainerStarted","Data":"dc0d55dcda6f99345c2422b205628947d98433492f615f9dbdd818c8817a1c4d"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.315719 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-n6cl2" event={"ID":"7b007c86-dbed-47a0-b456-aa6be1e6b21c","Type":"ContainerStarted","Data":"875ea16e858dc23e85a51c08b80310ff50a2bf557f20ddb7bdbf88b4818b42c2"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.317516 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5l7" event={"ID":"cbca6023-6b5b-43a6-aee8-c5df2797b649","Type":"ContainerStarted","Data":"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.317559 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5l7" event={"ID":"cbca6023-6b5b-43a6-aee8-c5df2797b649","Type":"ContainerStarted","Data":"d184b22acf5d0584d7ac378b2cca7fb221e52128099ea13db9539651d007ffad"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.323494 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" event={"ID":"7b85786e-8bab-4b2c-a908-977c8d0101c7","Type":"ContainerStarted","Data":"48609294a7162a1a883c89861d5bd9bd24c692dc86d71c77faf6556f34e1c5d0"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.383037 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.383313 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.883302009 +0000 UTC m=+122.233941961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.407467 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" event={"ID":"ea603264-026d-4e72-aabd-e7d85fe516d8","Type":"ContainerStarted","Data":"0c3c59d4987793bb96a49c5edf3485ecb7c6f5dc5aaebe320b3290c3fe988843"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.412892 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" event={"ID":"4a426294-bb9a-4532-a89e-36464ead5a80","Type":"ContainerStarted","Data":"03ca7aa972e8e357f16772c1b0ff6f1d87f629fbea8131ea14a3c73bf1bec010"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.456693 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" event={"ID":"e16847f2-730e-4ca5-a193-fde4db19cb2c","Type":"ContainerStarted","Data":"3c7d13f15238a3075e6c45d9e140bc17cbe74fc49d5fa3a6b6b148976c84b9ff"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.468358 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" event={"ID":"5cd140b8-9734-4f3f-947d-44937d49ffa1","Type":"ContainerStarted","Data":"789f7e91545ff7f942424cca92e1067e8d60cba41561ce5e842380236ecb6066"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.469378 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.470498 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" event={"ID":"41a72e46-e9a8-48f0-b002-b8930f2283a9","Type":"ContainerStarted","Data":"36493a002b0bfe36393d780ba0206a08fec618229a5d7dad8f6540a039d61941"} Oct 01 15:48:52 crc kubenswrapper[4688]: W1001 15:48:52.472033 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3f387ea_c091_4def_a25f_f299bb8ee3f9.slice/crio-74352b1c64868a142f34a284320dd05322cad97befde2f2c84981408ca95d842 WatchSource:0}: Error finding container 74352b1c64868a142f34a284320dd05322cad97befde2f2c84981408ca95d842: Status 404 returned error can't find the container with id 74352b1c64868a142f34a284320dd05322cad97befde2f2c84981408ca95d842 Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.474466 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v4x78" event={"ID":"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1","Type":"ContainerStarted","Data":"ed88acc420408cfcf659cd2a6bf1e89c85591a776ba7aff3f2c04500e0d701e9"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.476382 4688 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-f8thq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.476412 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" podUID="5cd140b8-9734-4f3f-947d-44937d49ffa1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.490447 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.493240 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.9932069 +0000 UTC m=+122.343846912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.493455 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.496436 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:52.996425745 +0000 UTC m=+122.347065707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.538943 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" event={"ID":"a1f921e5-85a7-4296-bb0d-587c064b48d9","Type":"ContainerStarted","Data":"f6b7b87ea061f1733304cf8b0e8b8e9b27b8460fbdacaa49f109084bf129b395"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.538992 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" event={"ID":"a1f921e5-85a7-4296-bb0d-587c064b48d9","Type":"ContainerStarted","Data":"d41e0e652a187cf85df3023e2fc527f7cfdb3e869f79efe4efbe4ac755ccd81d"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.542314 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.564559 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" event={"ID":"5de73e69-cb3e-4d2b-81b3-4ea56231536e","Type":"ContainerStarted","Data":"63edb2fad78df5ebce15dfd25ade4e3c3ed391c208ab542d61c84bed5c129ec1"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.576318 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" event={"ID":"2814f7d1-706b-450a-b47c-2ccc77ee6c20","Type":"ContainerStarted","Data":"bfa54583041b71a5f85e96d0e55f61982a2a9432a0c3d60821c88d2a6130cf08"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.593214 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" event={"ID":"2b776271-b6a4-42a2-b352-8b474016fb30","Type":"ContainerStarted","Data":"19b7d77739183622b88033b86d37f556b0e777d01641f8dba96738ab486dad4a"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.595864 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-992xd"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.596206 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.596909 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.09689563 +0000 UTC m=+122.447535592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.614723 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" event={"ID":"042bc31a-ee11-4740-baa9-52474d0e273f","Type":"ContainerStarted","Data":"9ec22bdf4c05f212ad2025c5303dd5fbbf1634461778080a69e74f06512d2c2d"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.697547 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.697829 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.197818598 +0000 UTC m=+122.548458560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.781164 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.798930 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.799110 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.299085126 +0000 UTC m=+122.649725078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.799298 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.799604 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.299597321 +0000 UTC m=+122.650237283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.831860 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" event={"ID":"2f53e292-fdef-4716-b217-0c4fab2dd64c","Type":"ContainerStarted","Data":"f33a272ff859d2f4f379fa94f662da8c4dcdf453f583de57b3dcc537a62bb6d4"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.831904 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" event={"ID":"2f53e292-fdef-4716-b217-0c4fab2dd64c","Type":"ContainerStarted","Data":"54daf4d96e9333bcab893a3245617a1acafbe248f2bef79553d9800e53f13632"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.858130 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.865733 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" podStartSLOduration=99.865711149 podStartE2EDuration="1m39.865711149s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:52.865327907 +0000 UTC m=+122.215967869" watchObservedRunningTime="2025-10-01 15:48:52.865711149 +0000 UTC m=+122.216351111" Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.877177 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pmzg7"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.886703 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2t92z"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.900505 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.900778 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.400711724 +0000 UTC m=+122.751351686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.901040 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:52 crc kubenswrapper[4688]: E1001 15:48:52.901372 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.401358933 +0000 UTC m=+122.751998895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.924921 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" event={"ID":"abb1e661-4852-4056-950f-90a2951e762e","Type":"ContainerStarted","Data":"b8310cffa6700ea70881248ec82083c22b95f46e2f7fd012277bb9ab8c757f95"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.948848 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x"] Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.952630 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-24qdq" event={"ID":"3fa56b51-5928-435f-887a-13adad0cf06e","Type":"ContainerStarted","Data":"66bfdb66e2c074bc3ceac6817adfc64980cae2ff9f0e151882d476a888dc4c45"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.961217 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" event={"ID":"4e28a890-8a72-42ff-9cff-7c3f100abdce","Type":"ContainerStarted","Data":"18a0351a0b623fb86ab1c2f1692dde410ee264794d57e92805969dc2723bfb07"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.980801 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" event={"ID":"8ca8a836-be9b-4c80-b8d5-43668ca982a5","Type":"ContainerStarted","Data":"613c5059a818080066e092ad90a2f32a82cba599879097a3074086101646995f"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.990751 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" event={"ID":"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f","Type":"ContainerStarted","Data":"707e7845f8058a0ee0e5a827aa7262e08a203528fdcbec968d444977fc22836a"} Oct 01 15:48:52 crc kubenswrapper[4688]: I1001 15:48:52.990784 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" event={"ID":"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f","Type":"ContainerStarted","Data":"6bb6a055e059b77302affc87baaa00fcae3a83d05899c1153623669411e109e5"} Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.001907 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.002670 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.502639092 +0000 UTC m=+122.853279054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.028173 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tt75q" event={"ID":"ca5f8204-18df-4b10-ab58-0ea02df49e4f","Type":"ContainerStarted","Data":"da8fc380a2a69d0fad5c71148e7a078c703011676cc98fe312421e1bbc9adf48"} Oct 01 15:48:53 crc kubenswrapper[4688]: W1001 15:48:53.040440 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5433040b_8666_4f49_98ce_9129ef301164.slice/crio-0086c57f36344be9c4adebaa3d855fd36c4ada9f3e75c67fc7d5656f22a975e9 WatchSource:0}: Error finding container 0086c57f36344be9c4adebaa3d855fd36c4ada9f3e75c67fc7d5656f22a975e9: Status 404 returned error can't find the container with id 0086c57f36344be9c4adebaa3d855fd36c4ada9f3e75c67fc7d5656f22a975e9 Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.041171 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" event={"ID":"4b66d47b-5d72-45ad-bb20-675829bf567a","Type":"ContainerStarted","Data":"d21beafe9688b1fae6a91aac072aebdce18d1eaf737d8a481d05625f60baa809"} Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.054729 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lx9hp"] Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.066318 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5"] Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.070766 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.078145 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs55m"] Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.078465 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.105637 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.105950 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.60593969 +0000 UTC m=+122.956579652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.216811 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.219367 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.719274122 +0000 UTC m=+123.069914124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.229956 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj"] Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.324082 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.324423 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.824409343 +0000 UTC m=+123.175049305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.399232 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mbmn8" podStartSLOduration=100.399219326 podStartE2EDuration="1m40.399219326s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.330886993 +0000 UTC m=+122.681526955" watchObservedRunningTime="2025-10-01 15:48:53.399219326 +0000 UTC m=+122.749859288" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.424970 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.425315 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:53.9252923 +0000 UTC m=+123.275932262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: W1001 15:48:53.436691 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13ad828f_f06e_48a3_a2ba_bb6f757b356e.slice/crio-242abcbc04207e6f8ceea94cd81555cbab87a10bae4464f9a1f681f2c829911f WatchSource:0}: Error finding container 242abcbc04207e6f8ceea94cd81555cbab87a10bae4464f9a1f681f2c829911f: Status 404 returned error can't find the container with id 242abcbc04207e6f8ceea94cd81555cbab87a10bae4464f9a1f681f2c829911f Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.452636 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8r5l7" podStartSLOduration=101.452621901 podStartE2EDuration="1m41.452621901s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.425486056 +0000 UTC m=+122.776126018" watchObservedRunningTime="2025-10-01 15:48:53.452621901 +0000 UTC m=+122.803261853" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.468272 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" podStartSLOduration=100.468257099 podStartE2EDuration="1m40.468257099s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.466397005 +0000 UTC m=+122.817036967" watchObservedRunningTime="2025-10-01 15:48:53.468257099 +0000 UTC m=+122.818897061" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.513236 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kdnf8" podStartSLOduration=100.513213767 podStartE2EDuration="1m40.513213767s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.491604684 +0000 UTC m=+122.842244646" watchObservedRunningTime="2025-10-01 15:48:53.513213767 +0000 UTC m=+122.863853729" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.526788 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.527178 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.027160706 +0000 UTC m=+123.377800668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.612322 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2tdnx" podStartSLOduration=101.612306351 podStartE2EDuration="1m41.612306351s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.610024645 +0000 UTC m=+122.960664607" watchObservedRunningTime="2025-10-01 15:48:53.612306351 +0000 UTC m=+122.962946313" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.627608 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.627830 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.127809716 +0000 UTC m=+123.478449678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.627870 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.628262 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.128256239 +0000 UTC m=+123.478896201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.652941 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dkwrm" podStartSLOduration=101.652925272 podStartE2EDuration="1m41.652925272s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:53.65183285 +0000 UTC m=+123.002472812" watchObservedRunningTime="2025-10-01 15:48:53.652925272 +0000 UTC m=+123.003565234" Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.731027 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.731363 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.231349411 +0000 UTC m=+123.581989373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.831715 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.832032 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.332020681 +0000 UTC m=+123.682660643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:53 crc kubenswrapper[4688]: I1001 15:48:53.932559 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:53 crc kubenswrapper[4688]: E1001 15:48:53.933161 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.433145145 +0000 UTC m=+123.783785107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.034685 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.035056 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.535044452 +0000 UTC m=+123.885684414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.060750 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" event={"ID":"d3f387ea-c091-4def-a25f-f299bb8ee3f9","Type":"ContainerStarted","Data":"74352b1c64868a142f34a284320dd05322cad97befde2f2c84981408ca95d842"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.078147 4688 generic.go:334] "Generic (PLEG): container finished" podID="41a72e46-e9a8-48f0-b002-b8930f2283a9" containerID="63238e2fde3d127120ba87609e05f507a096cac8116205771167b84c2fa9f0d9" exitCode=0 Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.078296 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" event={"ID":"41a72e46-e9a8-48f0-b002-b8930f2283a9","Type":"ContainerDied","Data":"63238e2fde3d127120ba87609e05f507a096cac8116205771167b84c2fa9f0d9"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.086105 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" event={"ID":"7c051410-74cd-4aad-9090-437673913d17","Type":"ContainerStarted","Data":"a9ee81390f04f50caebf1c18c98160ac04d0017503642ce3259ea20da5cfa8c6"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.087189 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" event={"ID":"ce79220d-2d22-4803-afcb-af02069a0f8b","Type":"ContainerStarted","Data":"ab989feb5b5e3d06e51893f52b6418b366d5b5e3b5bd6ad989346dd2120f558b"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.093596 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" event={"ID":"7b85786e-8bab-4b2c-a908-977c8d0101c7","Type":"ContainerStarted","Data":"83879c6321b9d2edf31a3220130101e8225de584ea5aedd4bbd1839cf2aad2fd"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.103233 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" event={"ID":"3d403b13-9b49-4193-9589-4f29e68ea9a6","Type":"ContainerStarted","Data":"aecf7618f51048c6666dcc95cc6f55cdddb81276b052414e74a8b9b6ad43dfb2"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.126647 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" event={"ID":"5433040b-8666-4f49-98ce-9129ef301164","Type":"ContainerStarted","Data":"0086c57f36344be9c4adebaa3d855fd36c4ada9f3e75c67fc7d5656f22a975e9"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.131077 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" event={"ID":"9b6bc4f2-3197-45e8-b934-c666e37c167e","Type":"ContainerStarted","Data":"634bd3d47930a39e128fce5f145d4313ad8a4025db729d43bcd46bc3ad2624ef"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.137065 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.138941 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.638921947 +0000 UTC m=+123.989561909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.151066 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" event={"ID":"b7ec1934-3a2e-4290-aa7e-581bbb58163e","Type":"ContainerStarted","Data":"449f694819399815db44a0076eac5d0cc16a9fa64e014ce59537056a19e633d6"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.160889 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" event={"ID":"2b776271-b6a4-42a2-b352-8b474016fb30","Type":"ContainerStarted","Data":"bc3dbb890b99bbb780cbb6658ace2d41a6867f8d544f001a56f5d80a18414fda"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.174022 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" event={"ID":"4e28a890-8a72-42ff-9cff-7c3f100abdce","Type":"ContainerStarted","Data":"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.176797 4688 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-m94rs container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" start-of-body= Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.176825 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": dial tcp 10.217.0.33:6443: connect: connection refused" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.176880 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.177822 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" event={"ID":"17f98bf7-aac6-4e31-8ebd-9237db04f612","Type":"ContainerStarted","Data":"15d420a5811ee96c8e569028b54643bcc90fb0a5726c67c960e6410664490775"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.178909 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" event={"ID":"13ad828f-f06e-48a3-a2ba-bb6f757b356e","Type":"ContainerStarted","Data":"242abcbc04207e6f8ceea94cd81555cbab87a10bae4464f9a1f681f2c829911f"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.204003 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5zwvb" podStartSLOduration=101.203974303 podStartE2EDuration="1m41.203974303s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.164564648 +0000 UTC m=+123.515204610" watchObservedRunningTime="2025-10-01 15:48:54.203974303 +0000 UTC m=+123.554614265" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.216771 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-24qdq" event={"ID":"3fa56b51-5928-435f-887a-13adad0cf06e","Type":"ContainerStarted","Data":"4ab8b676f10c4ca8a56d4a93955c4637abb554a44df021d70e1bdfd96d632b22"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.232730 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" event={"ID":"47fa404e-12b7-4590-8a04-7b304962b3ea","Type":"ContainerStarted","Data":"22d89356b23633efdf56a5d553202b90ff1fb2484245e6b6fb14057b7b20abeb"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.233988 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.237491 4688 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wzscd container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.237560 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" podUID="47fa404e-12b7-4590-8a04-7b304962b3ea" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.246445 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.248011 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.747996344 +0000 UTC m=+124.098636306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.249254 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" event={"ID":"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea","Type":"ContainerStarted","Data":"0b3731a88fa6cbddc8d65f75479870d0f8cd4ac229375ae872efdf38e01858a7"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.250707 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pmzg7" event={"ID":"dabdf9c3-fc17-46fd-bf40-5a67705823cc","Type":"ContainerStarted","Data":"27758669fe3887eb68e6b806693a373d0cb0d7ea51e2bcde1757ab008018eef0"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.294966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" event={"ID":"abb1e661-4852-4056-950f-90a2951e762e","Type":"ContainerStarted","Data":"927cb85fd46b661eca37e260e5f37d3cf6a9e8dffb1c034b4b5553690883292d"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.301943 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jrw8q" podStartSLOduration=101.301920024 podStartE2EDuration="1m41.301920024s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.204675304 +0000 UTC m=+123.555315266" watchObservedRunningTime="2025-10-01 15:48:54.301920024 +0000 UTC m=+123.652559986" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.333925 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" event={"ID":"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec","Type":"ContainerStarted","Data":"39f9bf86f96d9f791da26e45dfe7b900500d38a56a564bb109490d5d9dc62508"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.335103 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5sgb5" event={"ID":"ba2124c7-a29e-4a00-be83-9679fdbe8c7d","Type":"ContainerStarted","Data":"15b62bdbde7b487afc203d3593f46e7bee85f2992b3c9eb694eaedb2c2fccb3f"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.345283 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" event={"ID":"4a426294-bb9a-4532-a89e-36464ead5a80","Type":"ContainerStarted","Data":"e862e54524afe94fc67917f0d2067cefa38d8f448e5fe0a85fb73c533282fb1a"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.348168 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.349440 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.849419786 +0000 UTC m=+124.200059748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.362435 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" event={"ID":"4a8ec615-416d-4e20-b133-3adfc59692db","Type":"ContainerStarted","Data":"13fa47e29c6c157d95196f6722519641e03806778080247393723377e10f9f89"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.407073 4688 generic.go:334] "Generic (PLEG): container finished" podID="4b66d47b-5d72-45ad-bb20-675829bf567a" containerID="e127c5a0db83f4d1bc58bc760e8b26b58554b73e66275bd595bc2415aefa730f" exitCode=0 Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.407163 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" event={"ID":"4b66d47b-5d72-45ad-bb20-675829bf567a","Type":"ContainerDied","Data":"e127c5a0db83f4d1bc58bc760e8b26b58554b73e66275bd595bc2415aefa730f"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.417124 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rv2jj" podStartSLOduration=101.41710622 podStartE2EDuration="1m41.41710622s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.303117729 +0000 UTC m=+123.653757691" watchObservedRunningTime="2025-10-01 15:48:54.41710622 +0000 UTC m=+123.767746182" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.439749 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" event={"ID":"1228ad7f-0f56-4610-ba38-8cb9f64b4e0f","Type":"ContainerStarted","Data":"0fe774d955307605d497e169d4963fc9370e874fe17ba638076f1612965c7be7"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.449783 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.452488 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:54.952471227 +0000 UTC m=+124.303111239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.464114 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" event={"ID":"dc8ba611-f5be-48f7-990f-f56d4f7a9539","Type":"ContainerStarted","Data":"90b32402db483fc0298cb67515e5af4ada560ff0f521b4ee523ce7a1d9f45123"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.476729 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" event={"ID":"042bc31a-ee11-4740-baa9-52474d0e273f","Type":"ContainerStarted","Data":"477ef67dac0fa9b183cb4b1a0e13e15c6d70888ca623461737ee00b3394e6aaa"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.493331 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v4x78" event={"ID":"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1","Type":"ContainerStarted","Data":"172b4a5e31a6acb226fc9b318a8da45a2b279e48ee3e5a7c196efc1913f2a5d0"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.494988 4688 generic.go:334] "Generic (PLEG): container finished" podID="8ca8a836-be9b-4c80-b8d5-43668ca982a5" containerID="e6e59808b38f1e1409840961c6ba3f62497ef8e3082b13813f1c8a082c47d861" exitCode=0 Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.495077 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" event={"ID":"8ca8a836-be9b-4c80-b8d5-43668ca982a5","Type":"ContainerDied","Data":"e6e59808b38f1e1409840961c6ba3f62497ef8e3082b13813f1c8a082c47d861"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.551111 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.552255 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.052240781 +0000 UTC m=+124.402880743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.573250 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" event={"ID":"2814f7d1-706b-450a-b47c-2ccc77ee6c20","Type":"ContainerStarted","Data":"07540a13dc6a3d2b4db5cfb5e32abd72123c2a7791c324518d6c13a9853427d7"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.574875 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" podStartSLOduration=101.574854134 podStartE2EDuration="1m41.574854134s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.424066674 +0000 UTC m=+123.774706636" watchObservedRunningTime="2025-10-01 15:48:54.574854134 +0000 UTC m=+123.925494096" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.612404 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" event={"ID":"e16847f2-730e-4ca5-a193-fde4db19cb2c","Type":"ContainerStarted","Data":"0cac8c81d9fc46cd6cd4ad001d163a292856b134197d81631e84f64b986bb2ad"} Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.625909 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f8thq" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.653873 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.657615 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.157603159 +0000 UTC m=+124.508243121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.686847 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" podStartSLOduration=101.686830306 podStartE2EDuration="1m41.686830306s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.576329567 +0000 UTC m=+123.926969519" watchObservedRunningTime="2025-10-01 15:48:54.686830306 +0000 UTC m=+124.037470268" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.759125 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.759472 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.259456305 +0000 UTC m=+124.610096257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.809460 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-24qdq" podStartSLOduration=6.8094456 podStartE2EDuration="6.8094456s" podCreationTimestamp="2025-10-01 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.736226664 +0000 UTC m=+124.086866626" watchObservedRunningTime="2025-10-01 15:48:54.8094456 +0000 UTC m=+124.160085562" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.854254 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nkp4l" podStartSLOduration=101.854241653 podStartE2EDuration="1m41.854241653s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.809968805 +0000 UTC m=+124.160608777" watchObservedRunningTime="2025-10-01 15:48:54.854241653 +0000 UTC m=+124.204881615" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.865754 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.866031 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.366022088 +0000 UTC m=+124.716662050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.904188 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mhll6" podStartSLOduration=101.904171386 podStartE2EDuration="1m41.904171386s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.855706256 +0000 UTC m=+124.206346218" watchObservedRunningTime="2025-10-01 15:48:54.904171386 +0000 UTC m=+124.254811338" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.954656 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-npqrk" podStartSLOduration=101.954641046 podStartE2EDuration="1m41.954641046s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:54.953246735 +0000 UTC m=+124.303886697" watchObservedRunningTime="2025-10-01 15:48:54.954641046 +0000 UTC m=+124.305281008" Oct 01 15:48:54 crc kubenswrapper[4688]: I1001 15:48:54.967370 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:54 crc kubenswrapper[4688]: E1001 15:48:54.967649 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.467635937 +0000 UTC m=+124.818275899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.002700 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q2vld" podStartSLOduration=102.002684124 podStartE2EDuration="1m42.002684124s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.000445348 +0000 UTC m=+124.351085320" watchObservedRunningTime="2025-10-01 15:48:55.002684124 +0000 UTC m=+124.353324086" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.070988 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.071741 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.571729498 +0000 UTC m=+124.922369460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.112204 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" podStartSLOduration=102.112181912 podStartE2EDuration="1m42.112181912s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.067763931 +0000 UTC m=+124.418403893" watchObservedRunningTime="2025-10-01 15:48:55.112181912 +0000 UTC m=+124.462821874" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.154826 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.155678 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.163192 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.172249 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.172877 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.672855711 +0000 UTC m=+125.023495673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.184435 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.273541 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.273577 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.273607 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk9rl\" (UniqueName: \"kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.273667 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.273933 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.773921753 +0000 UTC m=+125.124561715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.371076 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.372332 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.377052 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.377302 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.377335 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.377371 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk9rl\" (UniqueName: \"kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.377637 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.877624272 +0000 UTC m=+125.228264234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.377977 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.378242 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.405836 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.471912 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk9rl\" (UniqueName: \"kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl\") pod \"certified-operators-gkhsz\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.493850 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.493929 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.493961 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.493979 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbxf4\" (UniqueName: \"kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.494302 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:55.994291362 +0000 UTC m=+125.344931324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.494357 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.535819 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.561094 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.562893 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.577226 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.595106 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.595390 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.595422 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2dtz\" (UniqueName: \"kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.595451 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.596185 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.596233 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbxf4\" (UniqueName: \"kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.596287 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.596437 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.096419095 +0000 UTC m=+125.447059057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.599325 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.599909 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.684320 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbxf4\" (UniqueName: \"kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4\") pod \"community-operators-vj4w9\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.699908 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.699972 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.699990 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2dtz\" (UniqueName: \"kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.700031 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.700267 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.200257919 +0000 UTC m=+125.550897881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.700577 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.700780 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.701404 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.719757 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-n6cl2" event={"ID":"7b007c86-dbed-47a0-b456-aa6be1e6b21c","Type":"ContainerStarted","Data":"6fd107793698b627c411070adde30814e5763e08c0fa683f33c2ec6136760c17"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.719807 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.721388 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.721462 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.722335 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.722384 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.732873 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" event={"ID":"d3f387ea-c091-4def-a25f-f299bb8ee3f9","Type":"ContainerStarted","Data":"f0b9a09e65d6bb238b16481b63dd8002b570113c6e05f8c73fcce12e57248ebd"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.732909 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" event={"ID":"d3f387ea-c091-4def-a25f-f299bb8ee3f9","Type":"ContainerStarted","Data":"0a1d71a9351776bc10ee5163549f8f2d4611f1e11a7d1fd66aee1f072c5ac4e3"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.746447 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" event={"ID":"3d403b13-9b49-4193-9589-4f29e68ea9a6","Type":"ContainerStarted","Data":"0281d682fbd347e56db50be0b5b60e7c08d1a9d59763ad29590340aaefb31caa"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.754227 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2dtz\" (UniqueName: \"kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz\") pod \"certified-operators-2s7vk\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.764058 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-n6cl2" podStartSLOduration=103.764041218 podStartE2EDuration="1m43.764041218s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.759956419 +0000 UTC m=+125.110596381" watchObservedRunningTime="2025-10-01 15:48:55.764041218 +0000 UTC m=+125.114681180" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.764658 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.765997 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9kj7x" event={"ID":"dc8ba611-f5be-48f7-990f-f56d4f7a9539","Type":"ContainerStarted","Data":"f7346e82062978e1765260aa99e24e52f2be45abf4c8c50507f7a4bb94c88178"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.801361 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.801574 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4llc\" (UniqueName: \"kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.801699 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.801761 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.802343 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.302329331 +0000 UTC m=+125.652969293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.811090 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" event={"ID":"9b6bc4f2-3197-45e8-b934-c666e37c167e","Type":"ContainerStarted","Data":"c8dbb87a9cf649b4e1621ea4c4e1ca23f351fc779baf9e552ac8135f03d5624f"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.811379 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" event={"ID":"9b6bc4f2-3197-45e8-b934-c666e37c167e","Type":"ContainerStarted","Data":"4d9b36083fe9f1d6ce38cbd0fb1a2d191fc68377290bb2e28ec9b96c431db31d"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.811615 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.831148 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sq75x" podStartSLOduration=102.831131125 podStartE2EDuration="1m42.831131125s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.822140691 +0000 UTC m=+125.172780653" watchObservedRunningTime="2025-10-01 15:48:55.831131125 +0000 UTC m=+125.181771087" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.836098 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" event={"ID":"2f53e292-fdef-4716-b217-0c4fab2dd64c","Type":"ContainerStarted","Data":"39cfcf3f7328af0e4b3e16f46f540ac927212aaeb0f62a1ee08003686836f221"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.851202 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tt75q" event={"ID":"ca5f8204-18df-4b10-ab58-0ea02df49e4f","Type":"ContainerStarted","Data":"73c5e0c05297b083d29c140bc5f8f779b6d8fc9fb68dfe862c5ccdc12c1c69b8"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.851964 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.853157 4688 patch_prober.go:28] interesting pod/console-operator-58897d9998-tt75q container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.853188 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-tt75q" podUID="ca5f8204-18df-4b10-ab58-0ea02df49e4f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.877138 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" event={"ID":"b7ec1934-3a2e-4290-aa7e-581bbb58163e","Type":"ContainerStarted","Data":"adbe432effa45d81db128c2109f68cc1d3218444aacbafdbcaceef21bdc47669"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.904254 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.904315 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.904356 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.904401 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4llc\" (UniqueName: \"kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: E1001 15:48:55.904772 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.404751523 +0000 UTC m=+125.755391555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.905904 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.906251 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.916636 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" podStartSLOduration=102.916620881 podStartE2EDuration="1m42.916620881s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.910975505 +0000 UTC m=+125.261615467" watchObservedRunningTime="2025-10-01 15:48:55.916620881 +0000 UTC m=+125.267260843" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.937182 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" event={"ID":"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea","Type":"ContainerStarted","Data":"a7082a1240ebc8e45b1f679790a3b3cceed89a3f90fef7c538a61fae6c904054"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.939559 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" event={"ID":"13ad828f-f06e-48a3-a2ba-bb6f757b356e","Type":"ContainerStarted","Data":"8d66ebd620ebdee5230f6684e9da734b137f7f04b569a70413522db150637350"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.940557 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.941582 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4llc\" (UniqueName: \"kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc\") pod \"community-operators-r5sw4\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.947574 4688 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wxhpj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.947604 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" podUID="13ad828f-f06e-48a3-a2ba-bb6f757b356e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.986042 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" event={"ID":"41a72e46-e9a8-48f0-b002-b8930f2283a9","Type":"ContainerStarted","Data":"bbd978574bd0b73db145939d56dbf1fa4fac4723292d98d4f1b1f203d5301f70"} Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.986664 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:55 crc kubenswrapper[4688]: I1001 15:48:55.986784 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.014561 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.014758 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-tt75q" podStartSLOduration=104.014741407 podStartE2EDuration="1m44.014741407s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:55.950617827 +0000 UTC m=+125.301257789" watchObservedRunningTime="2025-10-01 15:48:56.014741407 +0000 UTC m=+125.365381369" Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.014974 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.514957303 +0000 UTC m=+125.865597265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.015811 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5sgb5" event={"ID":"ba2124c7-a29e-4a00-be83-9679fdbe8c7d","Type":"ContainerStarted","Data":"722ac620373ee29790a491a859796e1bc32e8c8566cc6f0714eb19c6736e292b"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.047746 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" event={"ID":"17f98bf7-aac6-4e31-8ebd-9237db04f612","Type":"ContainerStarted","Data":"d38d88aea511c7436dda1b8c17b6502bb9c312452eccb9af9692fe1ad1bf28ae"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.051855 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.069498 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" podStartSLOduration=104.069479921 podStartE2EDuration="1m44.069479921s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.06400133 +0000 UTC m=+125.414641282" watchObservedRunningTime="2025-10-01 15:48:56.069479921 +0000 UTC m=+125.420119883" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.071007 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-szkjs" podStartSLOduration=103.071000605 podStartE2EDuration="1m43.071000605s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.014359605 +0000 UTC m=+125.364999567" watchObservedRunningTime="2025-10-01 15:48:56.071000605 +0000 UTC m=+125.421640567" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.071758 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" event={"ID":"8699a8ba-1dc6-49df-ae9b-d1feb4db64ec","Type":"ContainerStarted","Data":"539b251b0b5d085787ec343f9d0d0d512d9521e493a60e7d691277c97935a2b6"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.084398 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" event={"ID":"7c051410-74cd-4aad-9090-437673913d17","Type":"ContainerStarted","Data":"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.085236 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.091006 4688 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vfflp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.091042 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.108495 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" event={"ID":"e16847f2-730e-4ca5-a193-fde4db19cb2c","Type":"ContainerStarted","Data":"02dbe76a1ee581d3d03ea380f4b028ca581ae5174cfdaaa6654173c464aebc79"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.132598 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.132889 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pmzg7" event={"ID":"dabdf9c3-fc17-46fd-bf40-5a67705823cc","Type":"ContainerStarted","Data":"d6291d16c904896946a507c37f3f57a7a2d1ea32ad5d48b2335ca7e6c0c49a17"} Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.134837 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.634823586 +0000 UTC m=+125.985463548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.172683 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" event={"ID":"5433040b-8666-4f49-98ce-9129ef301164","Type":"ContainerStarted","Data":"c3c5162785e11e841cf6ed5a8d79bceb24dda42642378ff51434a3f729ce8c16"} Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.208838 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wzscd" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.244933 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.245005 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.744988945 +0000 UTC m=+126.095628907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.246038 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.263515 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.763500498 +0000 UTC m=+126.114140450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.309446 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" podStartSLOduration=103.309423214 podStartE2EDuration="1m43.309423214s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.21786532 +0000 UTC m=+125.568505282" watchObservedRunningTime="2025-10-01 15:48:56.309423214 +0000 UTC m=+125.660063176" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.348403 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.348785 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.848758917 +0000 UTC m=+126.199398879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.392785 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" podStartSLOduration=103.392768127 podStartE2EDuration="1m43.392768127s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.288893432 +0000 UTC m=+125.639533394" watchObservedRunningTime="2025-10-01 15:48:56.392768127 +0000 UTC m=+125.743408089" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.393976 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" podStartSLOduration=104.393970722 podStartE2EDuration="1m44.393970722s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.384600237 +0000 UTC m=+125.735240199" watchObservedRunningTime="2025-10-01 15:48:56.393970722 +0000 UTC m=+125.744610684" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.453395 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.453718 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:56.953707573 +0000 UTC m=+126.304347535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.555274 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.556065 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.056046042 +0000 UTC m=+126.406686004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.588660 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.596765 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:48:56 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:48:56 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:48:56 crc kubenswrapper[4688]: healthz check failed Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.596814 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.657763 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.658059 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.158049242 +0000 UTC m=+126.508689204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.705656 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lhmpv" podStartSLOduration=103.705641887 podStartE2EDuration="1m43.705641887s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.699760385 +0000 UTC m=+126.050400337" watchObservedRunningTime="2025-10-01 15:48:56.705641887 +0000 UTC m=+126.056281849" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.706440 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gmsfp" podStartSLOduration=103.70643554 podStartE2EDuration="1m43.70643554s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.5307396 +0000 UTC m=+125.881379552" watchObservedRunningTime="2025-10-01 15:48:56.70643554 +0000 UTC m=+126.057075502" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.763006 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.763382 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.263368479 +0000 UTC m=+126.614008441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.869597 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.870172 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.370161599 +0000 UTC m=+126.720801561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.939521 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nd6cx" podStartSLOduration=104.939498721 podStartE2EDuration="1m44.939498721s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:56.820924356 +0000 UTC m=+126.171564318" watchObservedRunningTime="2025-10-01 15:48:56.939498721 +0000 UTC m=+126.290138683" Oct 01 15:48:56 crc kubenswrapper[4688]: I1001 15:48:56.970504 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:56 crc kubenswrapper[4688]: E1001 15:48:56.970832 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.470816919 +0000 UTC m=+126.821456881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.012508 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.073648 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.073933 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.573923151 +0000 UTC m=+126.924563113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.082697 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-2t92z" podStartSLOduration=104.082682048 podStartE2EDuration="1m44.082682048s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.049077453 +0000 UTC m=+126.399717425" watchObservedRunningTime="2025-10-01 15:48:57.082682048 +0000 UTC m=+126.433322010" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.084895 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.176749 4688 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-m94rs container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.33:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.177094 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.33:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.177545 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.177904 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.677888729 +0000 UTC m=+127.028528691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.205644 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" podStartSLOduration=104.205626612 podStartE2EDuration="1m44.205626612s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.143304685 +0000 UTC m=+126.493944647" watchObservedRunningTime="2025-10-01 15:48:57.205626612 +0000 UTC m=+126.556266574" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.206223 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v4x78" event={"ID":"c8d1bcc4-727d-4e91-a6c6-3658cee8baa1","Type":"ContainerStarted","Data":"7807d73cdcdd8d310b9c7be323cbb1e492cbd301663d9c1ade113cdf12513bcd"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.206941 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-v4x78" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.221086 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pmzg7" podStartSLOduration=9.221072014 podStartE2EDuration="9.221072014s" podCreationTimestamp="2025-10-01 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.212930116 +0000 UTC m=+126.563570078" watchObservedRunningTime="2025-10-01 15:48:57.221072014 +0000 UTC m=+126.571711976" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.222332 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.241264 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" event={"ID":"4b66d47b-5d72-45ad-bb20-675829bf567a","Type":"ContainerStarted","Data":"252981b81c47d58104cc44150bed8307af34a522a0499ea99f405ead48a5901a"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.241307 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" event={"ID":"4b66d47b-5d72-45ad-bb20-675829bf567a","Type":"ContainerStarted","Data":"654026a6a857662c6bbbec6854b6b061b70b7892b16bd8b2b6c92c0155e83df7"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.270316 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerStarted","Data":"b6d698527955d0614613f9f54507065c03f6b5be3d89365f9851150983e74bb7"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.272898 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.280827 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.282008 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.78199498 +0000 UTC m=+127.132635022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.283765 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" event={"ID":"8ca8a836-be9b-4c80-b8d5-43668ca982a5","Type":"ContainerStarted","Data":"d5862d330c4508a64247525b5809447a6daac820dfb89ddc681a2ba9fe8abd78"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.303138 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" event={"ID":"042bc31a-ee11-4740-baa9-52474d0e273f","Type":"ContainerStarted","Data":"aeed6798387500d44a4d4df9e1fd60580982c4d35a7eb9a36ef5ccb046977fed"} Oct 01 15:48:57 crc kubenswrapper[4688]: W1001 15:48:57.304252 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc2ef8f0_df67_42bf_ad04_5361d0e48c71.slice/crio-70dbd5c0ff366ed34686ba85b8be542d4f401e65fb5a4f22e312cfa1e194e538 WatchSource:0}: Error finding container 70dbd5c0ff366ed34686ba85b8be542d4f401e65fb5a4f22e312cfa1e194e538: Status 404 returned error can't find the container with id 70dbd5c0ff366ed34686ba85b8be542d4f401e65fb5a4f22e312cfa1e194e538 Oct 01 15:48:57 crc kubenswrapper[4688]: W1001 15:48:57.307932 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod770f9d60_7d86_465b_9ef9_7c0ef8966c5d.slice/crio-afeba9525a2ebcc391044283f03b521bce9069221d2547c386caa62d61e1fe03 WatchSource:0}: Error finding container afeba9525a2ebcc391044283f03b521bce9069221d2547c386caa62d61e1fe03: Status 404 returned error can't find the container with id afeba9525a2ebcc391044283f03b521bce9069221d2547c386caa62d61e1fe03 Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.320804 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" event={"ID":"4a8ec615-416d-4e20-b133-3adfc59692db","Type":"ContainerStarted","Data":"8bafe5d471378086b935cd991a5cc084539d3d3d2a94b904fa2e2bbdd9ca202c"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.321207 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5sgb5" podStartSLOduration=104.321196319 podStartE2EDuration="1m44.321196319s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.307737064 +0000 UTC m=+126.658377026" watchObservedRunningTime="2025-10-01 15:48:57.321196319 +0000 UTC m=+126.671836271" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.323741 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.324624 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.337301 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.337634 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerStarted","Data":"17e03b5148bb12ceeee5e14b32b5af0863c2fd5bfb98ca173e4b5684223e8ba7"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.338359 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zzw65" podStartSLOduration=104.338345902 podStartE2EDuration="1m44.338345902s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.337931059 +0000 UTC m=+126.688571031" watchObservedRunningTime="2025-10-01 15:48:57.338345902 +0000 UTC m=+126.688985864" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.363758 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" event={"ID":"3d403b13-9b49-4193-9589-4f29e68ea9a6","Type":"ContainerStarted","Data":"4618c9a21ffcf79c4c39b2851dc1e5653a2eb4febec3aff1b63b8f62ed08c4e6"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.386146 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.386234 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.886219925 +0000 UTC m=+127.236859887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.386890 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.388372 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:57.888353427 +0000 UTC m=+127.238993449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.392541 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-v4x78" podStartSLOduration=9.392506259 podStartE2EDuration="9.392506259s" podCreationTimestamp="2025-10-01 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.375216902 +0000 UTC m=+126.725856864" watchObservedRunningTime="2025-10-01 15:48:57.392506259 +0000 UTC m=+126.743146221" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.397861 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.397941 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtgrx\" (UniqueName: \"kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.398011 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.399693 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.399726 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.412056 4688 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vfflp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.412098 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.482557 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" podStartSLOduration=104.482544168 podStartE2EDuration="1m44.482544168s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.475730078 +0000 UTC m=+126.826370040" watchObservedRunningTime="2025-10-01 15:48:57.482544168 +0000 UTC m=+126.833184120" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.500781 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.500921 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" event={"ID":"90a6c9b1-4631-4c7a-8c7c-28fc4845e4ea","Type":"ContainerStarted","Data":"b911f8d974f2b3c5ca1ea393205c9859b07d746767b8e05031b3f04b459eefb4"} Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.505344 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.506516 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.506739 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtgrx\" (UniqueName: \"kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.507066 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.508413 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.509274 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.009259561 +0000 UTC m=+127.359899523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.515765 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.518121 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.587430 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-992xd" podStartSLOduration=104.587411622 podStartE2EDuration="1m44.587411622s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.580241652 +0000 UTC m=+126.930881614" watchObservedRunningTime="2025-10-01 15:48:57.587411622 +0000 UTC m=+126.938051584" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.594913 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtgrx\" (UniqueName: \"kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx\") pod \"redhat-marketplace-kp29h\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.595324 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:48:57 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:48:57 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:48:57 crc kubenswrapper[4688]: healthz check failed Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.595683 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.612617 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.617429 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.117417091 +0000 UTC m=+127.468057053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.657011 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.658087 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.683954 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lx9hp" podStartSLOduration=104.683941491 podStartE2EDuration="1m44.683941491s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:48:57.683117977 +0000 UTC m=+127.033757939" watchObservedRunningTime="2025-10-01 15:48:57.683941491 +0000 UTC m=+127.034581453" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.710004 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.714934 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.715554 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.215539687 +0000 UTC m=+127.566179649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.816847 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnpms\" (UniqueName: \"kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.816904 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.816923 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.816966 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.817259 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.317246048 +0000 UTC m=+127.667886010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.847915 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.918101 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.918325 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.41829772 +0000 UTC m=+127.768937672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.918440 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.918569 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnpms\" (UniqueName: \"kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.918608 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.918629 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: E1001 15:48:57.919239 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.419228057 +0000 UTC m=+127.769868019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.919818 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.919989 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.993089 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnpms\" (UniqueName: \"kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms\") pod \"redhat-marketplace-fw4mk\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:57 crc kubenswrapper[4688]: I1001 15:48:57.999886 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.019971 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.020122 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.520101424 +0000 UTC m=+127.870741386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.020354 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.020662 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.52065067 +0000 UTC m=+127.871290632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.121370 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.121642 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.621616359 +0000 UTC m=+127.972256321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.121815 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.122113 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.622101544 +0000 UTC m=+127.972741506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.223460 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.224013 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.72399931 +0000 UTC m=+128.074639272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.325594 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.325861 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.825850366 +0000 UTC m=+128.176490328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.399845 4688 patch_prober.go:28] interesting pod/console-operator-58897d9998-tt75q container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.400090 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-tt75q" podUID="ca5f8204-18df-4b10-ab58-0ea02df49e4f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.408903 4688 generic.go:334] "Generic (PLEG): container finished" podID="ad746ee8-2f78-40b9-b980-286172f023d4" containerID="eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c" exitCode=0 Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.409516 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerDied","Data":"eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.413025 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.413699 4688 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wxhpj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.413740 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" podUID="13ad828f-f06e-48a3-a2ba-bb6f757b356e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.422766 4688 generic.go:334] "Generic (PLEG): container finished" podID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerID="b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10" exitCode=0 Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.422833 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerDied","Data":"b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.422859 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerStarted","Data":"afeba9525a2ebcc391044283f03b521bce9069221d2547c386caa62d61e1fe03"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.429766 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.430159 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:58.930145262 +0000 UTC m=+128.280785224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.440284 4688 generic.go:334] "Generic (PLEG): container finished" podID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerID="b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320" exitCode=0 Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.440329 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerDied","Data":"b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.440354 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerStarted","Data":"70dbd5c0ff366ed34686ba85b8be542d4f401e65fb5a4f22e312cfa1e194e538"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.444634 4688 generic.go:334] "Generic (PLEG): container finished" podID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerID="382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629" exitCode=0 Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.445826 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerDied","Data":"382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629"} Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.446112 4688 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vfflp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.446138 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.531060 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.531875 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.532165 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.032150042 +0000 UTC m=+128.382790004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.532748 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.534822 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.553217 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.610267 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:48:58 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:48:58 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:48:58 crc kubenswrapper[4688]: healthz check failed Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.610305 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.631871 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.632288 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7kcj\" (UniqueName: \"kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.632547 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.632583 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.652400 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.152381206 +0000 UTC m=+128.503021168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.734056 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7kcj\" (UniqueName: \"kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.734158 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.734181 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.734294 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.734596 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.234583875 +0000 UTC m=+128.585223837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.741851 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.745893 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.795371 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7kcj\" (UniqueName: \"kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj\") pod \"redhat-operators-kcmw2\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.836240 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.836655 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.336639366 +0000 UTC m=+128.687279328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.852970 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.921464 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.922424 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.936909 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.937388 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:58 crc kubenswrapper[4688]: E1001 15:48:58.937760 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.437747719 +0000 UTC m=+128.788387681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:58 crc kubenswrapper[4688]: I1001 15:48:58.990138 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.039906 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.040096 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.540075599 +0000 UTC m=+128.890715561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.040231 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.040336 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.040630 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.040683 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfljp\" (UniqueName: \"kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.040825 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.54081818 +0000 UTC m=+128.891458142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.051723 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.143359 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.143545 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.143580 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.143667 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfljp\" (UniqueName: \"kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.144032 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.644019415 +0000 UTC m=+128.994659377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.144366 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.144583 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.186314 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfljp\" (UniqueName: \"kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp\") pod \"redhat-operators-mxdmz\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.202424 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.203020 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.206864 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.207040 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.237639 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.244534 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.244888 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.744876531 +0000 UTC m=+129.095516493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.250571 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.346859 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.347164 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.347194 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.348213 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.84819858 +0000 UTC m=+129.198838542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.449074 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.449328 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.449348 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.450014 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:48:59.949998574 +0000 UTC m=+129.300638536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.450133 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.450159 4688 patch_prober.go:28] interesting pod/console-operator-58897d9998-tt75q container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.450183 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-tt75q" podUID="ca5f8204-18df-4b10-ab58-0ea02df49e4f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.450354 4688 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-tlmff container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.450367 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" podUID="41a72e46-e9a8-48f0-b002-b8930f2283a9" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.451788 4688 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wxhpj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.451847 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" podUID="13ad828f-f06e-48a3-a2ba-bb6f757b356e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.463037 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.513353 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.537129 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerStarted","Data":"f478b7294a733e4b61a3258c014d10dfdc7a56bcf4ab8be501421df944aa2ded"} Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.551361 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.551619 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.051590281 +0000 UTC m=+129.402230243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.551888 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.552147 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.052135757 +0000 UTC m=+129.402775719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.552330 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerStarted","Data":"6e193b024b0e1ad4642d0492e90bf74e71a3cc641af97923d863ec0fd04e4ae0"} Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.570974 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" event={"ID":"4a8ec615-416d-4e20-b133-3adfc59692db","Type":"ContainerStarted","Data":"3da5321947f7ebc0f4bf1f2185e042b3a8f2077265e0ea0c2e958cf53dcce3cc"} Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.602357 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:48:59 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:48:59 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:48:59 crc kubenswrapper[4688]: healthz check failed Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.602403 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.609302 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.653913 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.654816 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.154801756 +0000 UTC m=+129.505441718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.723118 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.755735 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.756008 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.255996452 +0000 UTC m=+129.606636414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.856712 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.857377 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.357361433 +0000 UTC m=+129.708001395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.880780 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tlmff" Oct 01 15:48:59 crc kubenswrapper[4688]: I1001 15:48:59.959005 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:48:59 crc kubenswrapper[4688]: E1001 15:48:59.959668 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.459650922 +0000 UTC m=+129.810290884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.059791 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.060285 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.56025937 +0000 UTC m=+129.910899332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.162262 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.162602 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.66258964 +0000 UTC m=+130.013229602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.263920 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.264236 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.764220939 +0000 UTC m=+130.114860901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.366550 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.366972 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.866945549 +0000 UTC m=+130.217585511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.390656 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.468176 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.468700 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:00.968664111 +0000 UTC m=+130.319304073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.528028 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.528071 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.529772 4688 patch_prober.go:28] interesting pod/console-f9d7485db-8r5l7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.529854 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8r5l7" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.571602 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.572679 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.072659629 +0000 UTC m=+130.423299591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.607002 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:00 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:00 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:00 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.607077 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.641888 4688 generic.go:334] "Generic (PLEG): container finished" podID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerID="70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3" exitCode=0 Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.642120 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerDied","Data":"70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.642202 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerStarted","Data":"b27f3491a54d0d0c59e6a5d84c6f95f3920682698412b85cefaa07164a0d183a"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.647677 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.647710 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.649104 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb","Type":"ContainerStarted","Data":"83b6cbd0044b8a4bb2966e0b9d936e20160649d710dde507b4779276d3ce7ebe"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.663705 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.672799 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.673077 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.173050651 +0000 UTC m=+130.523690623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.673237 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.674040 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.17402906 +0000 UTC m=+130.524669022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.688034 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" event={"ID":"4a8ec615-416d-4e20-b133-3adfc59692db","Type":"ContainerStarted","Data":"b91896be2826023ce98f836a9cad5c2b815b41442c94261f9bf4eec3aa2acd68"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.693376 4688 generic.go:334] "Generic (PLEG): container finished" podID="3f036344-df58-4ea8-b960-28908e1cac82" containerID="dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088" exitCode=0 Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.693614 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerDied","Data":"dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.739890 4688 generic.go:334] "Generic (PLEG): container finished" podID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerID="b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4" exitCode=0 Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.740279 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerDied","Data":"b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.774190 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.774493 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.274468614 +0000 UTC m=+130.625108576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.774616 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.776174 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.276161544 +0000 UTC m=+130.626801506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.824371 4688 generic.go:334] "Generic (PLEG): container finished" podID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerID="6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24" exitCode=0 Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.824416 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerDied","Data":"6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.824439 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerStarted","Data":"582257a138b757787f390a0cc1e2652d602c55340b8955559dcd4505c171e878"} Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.876272 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.876789 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.376759572 +0000 UTC m=+130.727399544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.881144 4688 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 01 15:49:00 crc kubenswrapper[4688]: I1001 15:49:00.977954 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:00 crc kubenswrapper[4688]: E1001 15:49:00.978380 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.47836189 +0000 UTC m=+130.829001852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.028316 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.028382 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.028369 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.028474 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.030489 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.030769 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.045128 4688 patch_prober.go:28] interesting pod/apiserver-76f77b778f-lkrfg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]log ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]etcd ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/generic-apiserver-start-informers ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/max-in-flight-filter ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 01 15:49:01 crc kubenswrapper[4688]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 01 15:49:01 crc kubenswrapper[4688]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/project.openshift.io-projectcache ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/openshift.io-startinformers ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 01 15:49:01 crc kubenswrapper[4688]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 01 15:49:01 crc kubenswrapper[4688]: livez check failed Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.045181 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" podUID="4b66d47b-5d72-45ad-bb20-675829bf567a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.079099 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.079402 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.579377951 +0000 UTC m=+130.930017923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.125282 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-tt75q" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.181148 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.182376 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.68236479 +0000 UTC m=+131.033004752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.282205 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.283299 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.783284598 +0000 UTC m=+131.133924560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.297571 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.383303 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.383708 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.883696241 +0000 UTC m=+131.234336203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.484627 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.484714 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.984410493 +0000 UTC m=+131.335050445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.485548 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: E1001 15:49:01.486428 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 15:49:01.986418952 +0000 UTC m=+131.337058904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-j7jkr" (UID: "761ff908-3503-4509-9fc7-b237a76e3262") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.544378 4688 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-01T15:49:00.881423089Z","Handler":null,"Name":""} Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.550986 4688 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.551020 4688 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.587303 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.587782 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wxhpj" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.588685 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.591151 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:01 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:01 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:01 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.591421 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.596257 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.690797 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.798459 4688 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.798496 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.856737 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" event={"ID":"4a8ec615-416d-4e20-b133-3adfc59692db","Type":"ContainerStarted","Data":"33d56e0298447b952f9c0f0661f4d9961b8c7e5392dfa63c19abd421ff980615"} Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.862716 4688 generic.go:334] "Generic (PLEG): container finished" podID="55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" containerID="275943df2d5a1fad7ceaa947e78323bdc16e4ffc8f657f29621a64a40624369b" exitCode=0 Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.864015 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb","Type":"ContainerDied","Data":"275943df2d5a1fad7ceaa947e78323bdc16e4ffc8f657f29621a64a40624369b"} Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.868814 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-j7jkr\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.881646 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zgxfs" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.890480 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-zs55m" podStartSLOduration=13.890465454 podStartE2EDuration="13.890465454s" podCreationTimestamp="2025-10-01 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:49:01.889854716 +0000 UTC m=+131.240494678" watchObservedRunningTime="2025-10-01 15:49:01.890465454 +0000 UTC m=+131.241105416" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.972412 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 15:49:01 crc kubenswrapper[4688]: I1001 15:49:01.980267 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.591816 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:02 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:02 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:02 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.592055 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.686094 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.942603 4688 generic.go:334] "Generic (PLEG): container finished" podID="b7ec1934-3a2e-4290-aa7e-581bbb58163e" containerID="adbe432effa45d81db128c2109f68cc1d3218444aacbafdbcaceef21bdc47669" exitCode=0 Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.942666 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" event={"ID":"b7ec1934-3a2e-4290-aa7e-581bbb58163e","Type":"ContainerDied","Data":"adbe432effa45d81db128c2109f68cc1d3218444aacbafdbcaceef21bdc47669"} Oct 01 15:49:02 crc kubenswrapper[4688]: I1001 15:49:02.968737 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" event={"ID":"761ff908-3503-4509-9fc7-b237a76e3262","Type":"ContainerStarted","Data":"5919a9742fa115faf19277dfb22a82634840a2b950883486239e7730ae90c0d4"} Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.410699 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.413815 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.527187 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir\") pod \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.527261 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access\") pod \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\" (UID: \"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb\") " Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.529495 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" (UID: "55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.536642 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" (UID: "55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.591411 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:03 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:03 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:03 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.591468 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.633141 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.633169 4688 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.987230 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb","Type":"ContainerDied","Data":"83b6cbd0044b8a4bb2966e0b9d936e20160649d710dde507b4779276d3ce7ebe"} Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.987270 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83b6cbd0044b8a4bb2966e0b9d936e20160649d710dde507b4779276d3ce7ebe" Oct 01 15:49:03 crc kubenswrapper[4688]: I1001 15:49:03.987323 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.008132 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" event={"ID":"761ff908-3503-4509-9fc7-b237a76e3262","Type":"ContainerStarted","Data":"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08"} Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.028569 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" podStartSLOduration=111.028551617 podStartE2EDuration="1m51.028551617s" podCreationTimestamp="2025-10-01 15:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:49:04.027742731 +0000 UTC m=+133.378382693" watchObservedRunningTime="2025-10-01 15:49:04.028551617 +0000 UTC m=+133.379191579" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.429635 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.548970 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume\") pod \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.549074 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfgtb\" (UniqueName: \"kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb\") pod \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.549162 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume\") pod \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\" (UID: \"b7ec1934-3a2e-4290-aa7e-581bbb58163e\") " Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.550013 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume" (OuterVolumeSpecName: "config-volume") pod "b7ec1934-3a2e-4290-aa7e-581bbb58163e" (UID: "b7ec1934-3a2e-4290-aa7e-581bbb58163e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.553200 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb" (OuterVolumeSpecName: "kube-api-access-tfgtb") pod "b7ec1934-3a2e-4290-aa7e-581bbb58163e" (UID: "b7ec1934-3a2e-4290-aa7e-581bbb58163e"). InnerVolumeSpecName "kube-api-access-tfgtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.558006 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b7ec1934-3a2e-4290-aa7e-581bbb58163e" (UID: "b7ec1934-3a2e-4290-aa7e-581bbb58163e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.597355 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:04 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:04 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:04 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.597410 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.650903 4688 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7ec1934-3a2e-4290-aa7e-581bbb58163e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.650932 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7ec1934-3a2e-4290-aa7e-581bbb58163e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:04 crc kubenswrapper[4688]: I1001 15:49:04.650941 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfgtb\" (UniqueName: \"kubernetes.io/projected/b7ec1934-3a2e-4290-aa7e-581bbb58163e-kube-api-access-tfgtb\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.057341 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.061564 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9" event={"ID":"b7ec1934-3a2e-4290-aa7e-581bbb58163e","Type":"ContainerDied","Data":"449f694819399815db44a0076eac5d0cc16a9fa64e014ce59537056a19e633d6"} Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.061589 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="449f694819399815db44a0076eac5d0cc16a9fa64e014ce59537056a19e633d6" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.061607 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.319687 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 15:49:05 crc kubenswrapper[4688]: E1001 15:49:05.325481 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" containerName="pruner" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.325509 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" containerName="pruner" Oct 01 15:49:05 crc kubenswrapper[4688]: E1001 15:49:05.325546 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7ec1934-3a2e-4290-aa7e-581bbb58163e" containerName="collect-profiles" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.325556 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7ec1934-3a2e-4290-aa7e-581bbb58163e" containerName="collect-profiles" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.326170 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="55c5fc6e-0926-4a16-ac8c-f2f7d529d2cb" containerName="pruner" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.326211 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7ec1934-3a2e-4290-aa7e-581bbb58163e" containerName="collect-profiles" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.326728 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.333548 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.335091 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.349581 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.462228 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.462298 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.564000 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.564058 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.564391 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.593911 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:05 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:05 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:05 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.593959 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.608687 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.690008 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:05 crc kubenswrapper[4688]: I1001 15:49:05.920385 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-v4x78" Oct 01 15:49:06 crc kubenswrapper[4688]: I1001 15:49:06.041886 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:49:06 crc kubenswrapper[4688]: I1001 15:49:06.068288 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-lkrfg" Oct 01 15:49:06 crc kubenswrapper[4688]: I1001 15:49:06.227180 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 15:49:06 crc kubenswrapper[4688]: W1001 15:49:06.260212 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod36e1f5bf_f45c_4801_8b48_ea87e4d57df8.slice/crio-bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9 WatchSource:0}: Error finding container bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9: Status 404 returned error can't find the container with id bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9 Oct 01 15:49:06 crc kubenswrapper[4688]: I1001 15:49:06.602781 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:06 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:06 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:06 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:06 crc kubenswrapper[4688]: I1001 15:49:06.603091 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:07 crc kubenswrapper[4688]: I1001 15:49:07.116459 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"36e1f5bf-f45c-4801-8b48-ea87e4d57df8","Type":"ContainerStarted","Data":"bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9"} Oct 01 15:49:07 crc kubenswrapper[4688]: I1001 15:49:07.591121 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:07 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:07 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:07 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:07 crc kubenswrapper[4688]: I1001 15:49:07.591356 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:08 crc kubenswrapper[4688]: I1001 15:49:08.589959 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:08 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:08 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:08 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:08 crc kubenswrapper[4688]: I1001 15:49:08.590038 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:09 crc kubenswrapper[4688]: I1001 15:49:09.158440 4688 generic.go:334] "Generic (PLEG): container finished" podID="36e1f5bf-f45c-4801-8b48-ea87e4d57df8" containerID="934ac2ea6f758da7a91b0caeafbf9941f740cfb9842514a39650c1f33b5fdc92" exitCode=0 Oct 01 15:49:09 crc kubenswrapper[4688]: I1001 15:49:09.158483 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"36e1f5bf-f45c-4801-8b48-ea87e4d57df8","Type":"ContainerDied","Data":"934ac2ea6f758da7a91b0caeafbf9941f740cfb9842514a39650c1f33b5fdc92"} Oct 01 15:49:09 crc kubenswrapper[4688]: I1001 15:49:09.598363 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:09 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:09 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:09 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:09 crc kubenswrapper[4688]: I1001 15:49:09.598420 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.485980 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.528235 4688 patch_prober.go:28] interesting pod/console-f9d7485db-8r5l7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.528293 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8r5l7" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.557321 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access\") pod \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.557489 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir\") pod \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\" (UID: \"36e1f5bf-f45c-4801-8b48-ea87e4d57df8\") " Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.557590 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "36e1f5bf-f45c-4801-8b48-ea87e4d57df8" (UID: "36e1f5bf-f45c-4801-8b48-ea87e4d57df8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.557837 4688 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.565250 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "36e1f5bf-f45c-4801-8b48-ea87e4d57df8" (UID: "36e1f5bf-f45c-4801-8b48-ea87e4d57df8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.590643 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:10 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:10 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:10 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.590694 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:10 crc kubenswrapper[4688]: I1001 15:49:10.659347 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36e1f5bf-f45c-4801-8b48-ea87e4d57df8-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.032399 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.032463 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.034768 4688 patch_prober.go:28] interesting pod/downloads-7954f5f757-n6cl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.034816 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n6cl2" podUID="7b007c86-dbed-47a0-b456-aa6be1e6b21c" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.170013 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"36e1f5bf-f45c-4801-8b48-ea87e4d57df8","Type":"ContainerDied","Data":"bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9"} Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.170047 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfcc8bf8b064b1ab2fad872e5d15c82438772964c7b125e71922552841d1f7e9" Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.170093 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.591208 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:11 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:11 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:11 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:11 crc kubenswrapper[4688]: I1001 15:49:11.591455 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:12 crc kubenswrapper[4688]: I1001 15:49:12.591779 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:12 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:12 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:12 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:12 crc kubenswrapper[4688]: I1001 15:49:12.591832 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:13 crc kubenswrapper[4688]: I1001 15:49:13.591609 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:13 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:13 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:13 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:13 crc kubenswrapper[4688]: I1001 15:49:13.591655 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:14 crc kubenswrapper[4688]: I1001 15:49:14.591363 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:14 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:14 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:14 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:14 crc kubenswrapper[4688]: I1001 15:49:14.591568 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:15 crc kubenswrapper[4688]: I1001 15:49:15.591990 4688 patch_prober.go:28] interesting pod/router-default-5444994796-5sgb5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 15:49:15 crc kubenswrapper[4688]: [-]has-synced failed: reason withheld Oct 01 15:49:15 crc kubenswrapper[4688]: [+]process-running ok Oct 01 15:49:15 crc kubenswrapper[4688]: healthz check failed Oct 01 15:49:15 crc kubenswrapper[4688]: I1001 15:49:15.592069 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5sgb5" podUID="ba2124c7-a29e-4a00-be83-9679fdbe8c7d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 15:49:16 crc kubenswrapper[4688]: I1001 15:49:16.116719 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:49:16 crc kubenswrapper[4688]: I1001 15:49:16.600560 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:49:16 crc kubenswrapper[4688]: I1001 15:49:16.603468 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5sgb5" Oct 01 15:49:20 crc kubenswrapper[4688]: I1001 15:49:20.563648 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:49:20 crc kubenswrapper[4688]: I1001 15:49:20.571134 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.034806 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-n6cl2" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.221865 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.221998 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.226344 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.226745 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.240511 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.323213 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.323284 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.325779 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.336373 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.341487 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.348358 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.350457 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.512871 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.531245 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.541171 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 15:49:21 crc kubenswrapper[4688]: I1001 15:49:21.988266 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:49:25 crc kubenswrapper[4688]: I1001 15:49:25.052656 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:49:25 crc kubenswrapper[4688]: I1001 15:49:25.052892 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:49:31 crc kubenswrapper[4688]: I1001 15:49:31.305122 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vxjr5" Oct 01 15:49:35 crc kubenswrapper[4688]: I1001 15:49:35.025903 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:49:35 crc kubenswrapper[4688]: I1001 15:49:35.028708 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 15:49:35 crc kubenswrapper[4688]: I1001 15:49:35.045105 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c987bdd3-e8bc-473e-99b2-61a143a95543-metrics-certs\") pod \"network-metrics-daemon-5sm44\" (UID: \"c987bdd3-e8bc-473e-99b2-61a143a95543\") " pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:49:35 crc kubenswrapper[4688]: I1001 15:49:35.049587 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 15:49:35 crc kubenswrapper[4688]: I1001 15:49:35.058087 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5sm44" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.177311 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.177874 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.178039 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x2dtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2s7vk_openshift-marketplace(bc2ef8f0-df67-42bf-ad04-5361d0e48c71): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.178054 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pk9rl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gkhsz_openshift-marketplace(e1cb0316-bcd2-46ad-8f23-db3c545eff1c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.179197 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gkhsz" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" Oct 01 15:49:40 crc kubenswrapper[4688]: E1001 15:49:40.180115 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2s7vk" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" Oct 01 15:49:41 crc kubenswrapper[4688]: E1001 15:49:41.384581 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2s7vk" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" Oct 01 15:49:41 crc kubenswrapper[4688]: E1001 15:49:41.384681 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gkhsz" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" Oct 01 15:49:41 crc kubenswrapper[4688]: E1001 15:49:41.448953 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 15:49:41 crc kubenswrapper[4688]: E1001 15:49:41.449185 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h4llc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-r5sw4_openshift-marketplace(770f9d60-7d86-465b-9ef9-7c0ef8966c5d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:41 crc kubenswrapper[4688]: E1001 15:49:41.451490 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-r5sw4" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.051358 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.051524 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dnpms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fw4mk_openshift-marketplace(34de0946-abf9-41fd-a0dd-38bf4febbee8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.052816 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fw4mk" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.054808 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.054937 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xtgrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kp29h_openshift-marketplace(3f036344-df58-4ea8-b960-28908e1cac82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.056103 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kp29h" podUID="3f036344-df58-4ea8-b960-28908e1cac82" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.096022 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.096141 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dbxf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vj4w9_openshift-marketplace(ad746ee8-2f78-40b9-b980-286172f023d4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:42 crc kubenswrapper[4688]: E1001 15:49:42.097412 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vj4w9" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.402959 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kp29h" podUID="3f036344-df58-4ea8-b960-28908e1cac82" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.403004 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vj4w9" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.403218 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fw4mk" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.403248 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-r5sw4" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.579670 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.580048 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7kcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kcmw2_openshift-marketplace(d54684b4-efba-4ebf-91bf-524247aafc4e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.581192 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kcmw2" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.617749 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.617886 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfljp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mxdmz_openshift-marketplace(78b53489-1e16-4490-a8a7-b739ea9bc090): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 15:49:45 crc kubenswrapper[4688]: E1001 15:49:45.619093 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mxdmz" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" Oct 01 15:49:45 crc kubenswrapper[4688]: I1001 15:49:45.918251 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5sm44"] Oct 01 15:49:45 crc kubenswrapper[4688]: W1001 15:49:45.926353 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc987bdd3_e8bc_473e_99b2_61a143a95543.slice/crio-0d73cb23a294ebbefe65d2c87a235955170d261f24d7237f338cf093b91355ab WatchSource:0}: Error finding container 0d73cb23a294ebbefe65d2c87a235955170d261f24d7237f338cf093b91355ab: Status 404 returned error can't find the container with id 0d73cb23a294ebbefe65d2c87a235955170d261f24d7237f338cf093b91355ab Oct 01 15:49:46 crc kubenswrapper[4688]: W1001 15:49:46.060871 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-7871d1ed817c983f7611fb53af9d6fbbdc039b580321b4588ab3a9e30203d81c WatchSource:0}: Error finding container 7871d1ed817c983f7611fb53af9d6fbbdc039b580321b4588ab3a9e30203d81c: Status 404 returned error can't find the container with id 7871d1ed817c983f7611fb53af9d6fbbdc039b580321b4588ab3a9e30203d81c Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.370448 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bd694cfb8fefbcc4e874bde7274b89fca9f65b374b892e476620b2688ec207b3"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.370510 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7871d1ed817c983f7611fb53af9d6fbbdc039b580321b4588ab3a9e30203d81c"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.374083 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7750761723e51c2fc5d419e96031036c81762cccd134a939f61259b8ae00b663"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.374160 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5d4d64d9f08f4b1deb8515a57606ee404d997145288d58ef9e8fc6037373e313"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.374487 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.377773 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5sm44" event={"ID":"c987bdd3-e8bc-473e-99b2-61a143a95543","Type":"ContainerStarted","Data":"b899f523a0a0b7fd085dbdb89446c4f80a226fe84e150327df2d3503cc279de3"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.377840 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5sm44" event={"ID":"c987bdd3-e8bc-473e-99b2-61a143a95543","Type":"ContainerStarted","Data":"0d73cb23a294ebbefe65d2c87a235955170d261f24d7237f338cf093b91355ab"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.380469 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4f55d5a3e2dddf2c5b6103bd7040dc16a525b42f3a550dc75113000fc59a1d11"} Oct 01 15:49:46 crc kubenswrapper[4688]: I1001 15:49:46.380504 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3214b8b568e309c4d45a56c4440f3e41b08895bb9691814c7470cce177637e0c"} Oct 01 15:49:46 crc kubenswrapper[4688]: E1001 15:49:46.381886 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kcmw2" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" Oct 01 15:49:46 crc kubenswrapper[4688]: E1001 15:49:46.382601 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mxdmz" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" Oct 01 15:49:47 crc kubenswrapper[4688]: I1001 15:49:47.391271 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5sm44" event={"ID":"c987bdd3-e8bc-473e-99b2-61a143a95543","Type":"ContainerStarted","Data":"f841788c6f3162a4ff29d19cc2dee26bee66325c950103d7c0b1cc9f5e2fd8a8"} Oct 01 15:49:47 crc kubenswrapper[4688]: I1001 15:49:47.413779 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-5sm44" podStartSLOduration=155.413760881 podStartE2EDuration="2m35.413760881s" podCreationTimestamp="2025-10-01 15:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:49:47.413664548 +0000 UTC m=+176.764304550" watchObservedRunningTime="2025-10-01 15:49:47.413760881 +0000 UTC m=+176.764400853" Oct 01 15:49:55 crc kubenswrapper[4688]: I1001 15:49:55.053328 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:49:55 crc kubenswrapper[4688]: I1001 15:49:55.053901 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:49:56 crc kubenswrapper[4688]: I1001 15:49:56.451653 4688 generic.go:334] "Generic (PLEG): container finished" podID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerID="ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a" exitCode=0 Oct 01 15:49:56 crc kubenswrapper[4688]: I1001 15:49:56.451762 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerDied","Data":"ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a"} Oct 01 15:49:58 crc kubenswrapper[4688]: I1001 15:49:58.471951 4688 generic.go:334] "Generic (PLEG): container finished" podID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerID="df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233" exitCode=0 Oct 01 15:49:58 crc kubenswrapper[4688]: I1001 15:49:58.472045 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerDied","Data":"df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233"} Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.481777 4688 generic.go:334] "Generic (PLEG): container finished" podID="3f036344-df58-4ea8-b960-28908e1cac82" containerID="23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe" exitCode=0 Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.481938 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerDied","Data":"23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe"} Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.486405 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerStarted","Data":"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8"} Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.492560 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerStarted","Data":"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164"} Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.527227 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2s7vk" podStartSLOduration=3.996137557 podStartE2EDuration="1m4.527199696s" podCreationTimestamp="2025-10-01 15:48:55 +0000 UTC" firstStartedPulling="2025-10-01 15:48:58.462043117 +0000 UTC m=+127.812683079" lastFinishedPulling="2025-10-01 15:49:58.993105236 +0000 UTC m=+188.343745218" observedRunningTime="2025-10-01 15:49:59.522161582 +0000 UTC m=+188.872801564" watchObservedRunningTime="2025-10-01 15:49:59.527199696 +0000 UTC m=+188.877839698" Oct 01 15:49:59 crc kubenswrapper[4688]: I1001 15:49:59.537778 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gkhsz" podStartSLOduration=4.638158217 podStartE2EDuration="1m4.537732107s" podCreationTimestamp="2025-10-01 15:48:55 +0000 UTC" firstStartedPulling="2025-10-01 15:48:58.462238383 +0000 UTC m=+127.812878345" lastFinishedPulling="2025-10-01 15:49:58.361812233 +0000 UTC m=+187.712452235" observedRunningTime="2025-10-01 15:49:59.536859932 +0000 UTC m=+188.887499934" watchObservedRunningTime="2025-10-01 15:49:59.537732107 +0000 UTC m=+188.888372079" Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.498595 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerStarted","Data":"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2"} Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.500362 4688 generic.go:334] "Generic (PLEG): container finished" podID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerID="90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493" exitCode=0 Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.500410 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerDied","Data":"90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493"} Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.502569 4688 generic.go:334] "Generic (PLEG): container finished" podID="ad746ee8-2f78-40b9-b980-286172f023d4" containerID="8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7" exitCode=0 Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.502603 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerDied","Data":"8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7"} Oct 01 15:50:00 crc kubenswrapper[4688]: I1001 15:50:00.536461 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kp29h" podStartSLOduration=4.363646288 podStartE2EDuration="1m3.536445663s" podCreationTimestamp="2025-10-01 15:48:57 +0000 UTC" firstStartedPulling="2025-10-01 15:49:00.696431937 +0000 UTC m=+130.047071899" lastFinishedPulling="2025-10-01 15:49:59.869231312 +0000 UTC m=+189.219871274" observedRunningTime="2025-10-01 15:50:00.533939221 +0000 UTC m=+189.884579183" watchObservedRunningTime="2025-10-01 15:50:00.536445663 +0000 UTC m=+189.887085625" Oct 01 15:50:01 crc kubenswrapper[4688]: I1001 15:50:01.516676 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerStarted","Data":"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3"} Oct 01 15:50:01 crc kubenswrapper[4688]: I1001 15:50:01.518953 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerStarted","Data":"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad"} Oct 01 15:50:01 crc kubenswrapper[4688]: I1001 15:50:01.523815 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerStarted","Data":"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e"} Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.535731 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerStarted","Data":"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb"} Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.537257 4688 generic.go:334] "Generic (PLEG): container finished" podID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerID="8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6" exitCode=0 Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.537297 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerDied","Data":"8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6"} Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.541085 4688 generic.go:334] "Generic (PLEG): container finished" podID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerID="1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad" exitCode=0 Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.541966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerDied","Data":"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad"} Oct 01 15:50:02 crc kubenswrapper[4688]: I1001 15:50:02.638269 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vj4w9" podStartSLOduration=5.119984751 podStartE2EDuration="1m7.63825529s" podCreationTimestamp="2025-10-01 15:48:55 +0000 UTC" firstStartedPulling="2025-10-01 15:48:58.412791214 +0000 UTC m=+127.763431166" lastFinishedPulling="2025-10-01 15:50:00.931061743 +0000 UTC m=+190.281701705" observedRunningTime="2025-10-01 15:50:02.618993199 +0000 UTC m=+191.969633161" watchObservedRunningTime="2025-10-01 15:50:02.63825529 +0000 UTC m=+191.988895252" Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.549587 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerStarted","Data":"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af"} Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.551901 4688 generic.go:334] "Generic (PLEG): container finished" podID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerID="a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb" exitCode=0 Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.551950 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerDied","Data":"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb"} Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.554699 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerStarted","Data":"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76"} Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.569888 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fw4mk" podStartSLOduration=6.314841771 podStartE2EDuration="1m6.569867494s" podCreationTimestamp="2025-10-01 15:48:57 +0000 UTC" firstStartedPulling="2025-10-01 15:49:00.750395678 +0000 UTC m=+130.101035630" lastFinishedPulling="2025-10-01 15:50:01.005421381 +0000 UTC m=+190.356061353" observedRunningTime="2025-10-01 15:50:02.640741811 +0000 UTC m=+191.991381773" watchObservedRunningTime="2025-10-01 15:50:03.569867494 +0000 UTC m=+192.920507476" Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.570086 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kcmw2" podStartSLOduration=3.190955876 podStartE2EDuration="1m5.5700784s" podCreationTimestamp="2025-10-01 15:48:58 +0000 UTC" firstStartedPulling="2025-10-01 15:49:00.655701273 +0000 UTC m=+130.006341235" lastFinishedPulling="2025-10-01 15:50:03.034823787 +0000 UTC m=+192.385463759" observedRunningTime="2025-10-01 15:50:03.56830367 +0000 UTC m=+192.918943632" watchObservedRunningTime="2025-10-01 15:50:03.5700784 +0000 UTC m=+192.920718382" Oct 01 15:50:03 crc kubenswrapper[4688]: I1001 15:50:03.624385 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r5sw4" podStartSLOduration=4.092374062 podStartE2EDuration="1m8.624372394s" podCreationTimestamp="2025-10-01 15:48:55 +0000 UTC" firstStartedPulling="2025-10-01 15:48:58.461706688 +0000 UTC m=+127.812346650" lastFinishedPulling="2025-10-01 15:50:02.99370502 +0000 UTC m=+192.344344982" observedRunningTime="2025-10-01 15:50:03.623088257 +0000 UTC m=+192.973728219" watchObservedRunningTime="2025-10-01 15:50:03.624372394 +0000 UTC m=+192.975012346" Oct 01 15:50:04 crc kubenswrapper[4688]: I1001 15:50:04.560855 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerStarted","Data":"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527"} Oct 01 15:50:04 crc kubenswrapper[4688]: I1001 15:50:04.582236 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mxdmz" podStartSLOduration=3.299874241 podStartE2EDuration="1m6.582202379s" podCreationTimestamp="2025-10-01 15:48:58 +0000 UTC" firstStartedPulling="2025-10-01 15:49:00.838548512 +0000 UTC m=+130.189188474" lastFinishedPulling="2025-10-01 15:50:04.12087665 +0000 UTC m=+193.471516612" observedRunningTime="2025-10-01 15:50:04.582122747 +0000 UTC m=+193.932762719" watchObservedRunningTime="2025-10-01 15:50:04.582202379 +0000 UTC m=+193.932842341" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.537439 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.537732 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.702048 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.702104 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.987968 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:05 crc kubenswrapper[4688]: I1001 15:50:05.988328 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.052774 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.052814 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.058277 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.065782 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.066039 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.114834 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.131299 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.606727 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:06 crc kubenswrapper[4688]: I1001 15:50:06.611397 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:07 crc kubenswrapper[4688]: I1001 15:50:07.848744 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:07 crc kubenswrapper[4688]: I1001 15:50:07.850169 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:07 crc kubenswrapper[4688]: I1001 15:50:07.908110 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.000941 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.001023 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.046072 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.621341 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.621745 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.855066 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.856005 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:08 crc kubenswrapper[4688]: I1001 15:50:08.902420 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.161941 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.162294 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2s7vk" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="registry-server" containerID="cri-o://42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8" gracePeriod=2 Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.251203 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.251266 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.303825 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.630662 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:09 crc kubenswrapper[4688]: I1001 15:50:09.635802 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.440016 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.592314 4688 generic.go:334] "Generic (PLEG): container finished" podID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerID="42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8" exitCode=0 Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.592370 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2s7vk" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.592412 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerDied","Data":"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8"} Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.592496 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2s7vk" event={"ID":"bc2ef8f0-df67-42bf-ad04-5361d0e48c71","Type":"ContainerDied","Data":"70dbd5c0ff366ed34686ba85b8be542d4f401e65fb5a4f22e312cfa1e194e538"} Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.592551 4688 scope.go:117] "RemoveContainer" containerID="42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.599332 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities\") pod \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.599464 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content\") pod \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.599784 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2dtz\" (UniqueName: \"kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz\") pod \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\" (UID: \"bc2ef8f0-df67-42bf-ad04-5361d0e48c71\") " Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.601203 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities" (OuterVolumeSpecName: "utilities") pod "bc2ef8f0-df67-42bf-ad04-5361d0e48c71" (UID: "bc2ef8f0-df67-42bf-ad04-5361d0e48c71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.605022 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz" (OuterVolumeSpecName: "kube-api-access-x2dtz") pod "bc2ef8f0-df67-42bf-ad04-5361d0e48c71" (UID: "bc2ef8f0-df67-42bf-ad04-5361d0e48c71"). InnerVolumeSpecName "kube-api-access-x2dtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.613207 4688 scope.go:117] "RemoveContainer" containerID="df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.633959 4688 scope.go:117] "RemoveContainer" containerID="b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.645945 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc2ef8f0-df67-42bf-ad04-5361d0e48c71" (UID: "bc2ef8f0-df67-42bf-ad04-5361d0e48c71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.656107 4688 scope.go:117] "RemoveContainer" containerID="42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8" Oct 01 15:50:10 crc kubenswrapper[4688]: E1001 15:50:10.656554 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8\": container with ID starting with 42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8 not found: ID does not exist" containerID="42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.656595 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8"} err="failed to get container status \"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8\": rpc error: code = NotFound desc = could not find container \"42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8\": container with ID starting with 42bb6bd2f72135e1ba28d335fe04ce674bcb11cad4885582e1848296f0476ae8 not found: ID does not exist" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.656658 4688 scope.go:117] "RemoveContainer" containerID="df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233" Oct 01 15:50:10 crc kubenswrapper[4688]: E1001 15:50:10.657162 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233\": container with ID starting with df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233 not found: ID does not exist" containerID="df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.657195 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233"} err="failed to get container status \"df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233\": rpc error: code = NotFound desc = could not find container \"df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233\": container with ID starting with df8abe1b6cbf1591847de83b5c03cca1de6333fa87cd13c9297d6c591a82f233 not found: ID does not exist" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.657217 4688 scope.go:117] "RemoveContainer" containerID="b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320" Oct 01 15:50:10 crc kubenswrapper[4688]: E1001 15:50:10.657452 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320\": container with ID starting with b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320 not found: ID does not exist" containerID="b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.657477 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320"} err="failed to get container status \"b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320\": rpc error: code = NotFound desc = could not find container \"b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320\": container with ID starting with b82b27714ba2fbb553af4b62fd703bb2802f30b573c27c1fb3e0ee4686d4f320 not found: ID does not exist" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.701565 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.701744 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2dtz\" (UniqueName: \"kubernetes.io/projected/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-kube-api-access-x2dtz\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.701791 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc2ef8f0-df67-42bf-ad04-5361d0e48c71-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.919472 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:50:10 crc kubenswrapper[4688]: I1001 15:50:10.925072 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2s7vk"] Oct 01 15:50:11 crc kubenswrapper[4688]: I1001 15:50:11.388936 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" path="/var/lib/kubelet/pods/bc2ef8f0-df67-42bf-ad04-5361d0e48c71/volumes" Oct 01 15:50:11 crc kubenswrapper[4688]: I1001 15:50:11.555787 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:50:11 crc kubenswrapper[4688]: I1001 15:50:11.556033 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fw4mk" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="registry-server" containerID="cri-o://aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e" gracePeriod=2 Oct 01 15:50:11 crc kubenswrapper[4688]: I1001 15:50:11.988282 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.081129 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.119128 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content\") pod \"34de0946-abf9-41fd-a0dd-38bf4febbee8\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.119186 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnpms\" (UniqueName: \"kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms\") pod \"34de0946-abf9-41fd-a0dd-38bf4febbee8\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.119989 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities\") pod \"34de0946-abf9-41fd-a0dd-38bf4febbee8\" (UID: \"34de0946-abf9-41fd-a0dd-38bf4febbee8\") " Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.120536 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities" (OuterVolumeSpecName: "utilities") pod "34de0946-abf9-41fd-a0dd-38bf4febbee8" (UID: "34de0946-abf9-41fd-a0dd-38bf4febbee8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.126248 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms" (OuterVolumeSpecName: "kube-api-access-dnpms") pod "34de0946-abf9-41fd-a0dd-38bf4febbee8" (UID: "34de0946-abf9-41fd-a0dd-38bf4febbee8"). InnerVolumeSpecName "kube-api-access-dnpms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.138784 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34de0946-abf9-41fd-a0dd-38bf4febbee8" (UID: "34de0946-abf9-41fd-a0dd-38bf4febbee8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.221596 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.221631 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnpms\" (UniqueName: \"kubernetes.io/projected/34de0946-abf9-41fd-a0dd-38bf4febbee8-kube-api-access-dnpms\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.221645 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34de0946-abf9-41fd-a0dd-38bf4febbee8-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.558904 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.559097 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mxdmz" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="registry-server" containerID="cri-o://1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527" gracePeriod=2 Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.604902 4688 generic.go:334] "Generic (PLEG): container finished" podID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerID="aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e" exitCode=0 Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.604962 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fw4mk" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.604953 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerDied","Data":"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e"} Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.605081 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fw4mk" event={"ID":"34de0946-abf9-41fd-a0dd-38bf4febbee8","Type":"ContainerDied","Data":"6e193b024b0e1ad4642d0492e90bf74e71a3cc641af97923d863ec0fd04e4ae0"} Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.605100 4688 scope.go:117] "RemoveContainer" containerID="aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.618502 4688 scope.go:117] "RemoveContainer" containerID="90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.665698 4688 scope.go:117] "RemoveContainer" containerID="b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.675157 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.706681 4688 scope.go:117] "RemoveContainer" containerID="aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.710409 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fw4mk"] Oct 01 15:50:12 crc kubenswrapper[4688]: E1001 15:50:12.715565 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e\": container with ID starting with aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e not found: ID does not exist" containerID="aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.715614 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e"} err="failed to get container status \"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e\": rpc error: code = NotFound desc = could not find container \"aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e\": container with ID starting with aedcd5844b61df2ebedf18770818412b07ab6af78ca7d07c5baaa105a2f4c71e not found: ID does not exist" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.715639 4688 scope.go:117] "RemoveContainer" containerID="90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493" Oct 01 15:50:12 crc kubenswrapper[4688]: E1001 15:50:12.716083 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493\": container with ID starting with 90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493 not found: ID does not exist" containerID="90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.716104 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493"} err="failed to get container status \"90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493\": rpc error: code = NotFound desc = could not find container \"90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493\": container with ID starting with 90fe19ac165d311fd703183723e5efc81a5190b53f4b85e6013d6435032dd493 not found: ID does not exist" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.716117 4688 scope.go:117] "RemoveContainer" containerID="b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4" Oct 01 15:50:12 crc kubenswrapper[4688]: E1001 15:50:12.717450 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4\": container with ID starting with b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4 not found: ID does not exist" containerID="b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.717487 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4"} err="failed to get container status \"b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4\": rpc error: code = NotFound desc = could not find container \"b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4\": container with ID starting with b6aa2ea2a363b710d4f04ec1c14b03d8281af6079b925d2bb426dd2bdd9ceac4 not found: ID does not exist" Oct 01 15:50:12 crc kubenswrapper[4688]: I1001 15:50:12.980788 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.134878 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfljp\" (UniqueName: \"kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp\") pod \"78b53489-1e16-4490-a8a7-b739ea9bc090\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.134923 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities\") pod \"78b53489-1e16-4490-a8a7-b739ea9bc090\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.135063 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content\") pod \"78b53489-1e16-4490-a8a7-b739ea9bc090\" (UID: \"78b53489-1e16-4490-a8a7-b739ea9bc090\") " Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.138417 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp" (OuterVolumeSpecName: "kube-api-access-rfljp") pod "78b53489-1e16-4490-a8a7-b739ea9bc090" (UID: "78b53489-1e16-4490-a8a7-b739ea9bc090"). InnerVolumeSpecName "kube-api-access-rfljp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.142336 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities" (OuterVolumeSpecName: "utilities") pod "78b53489-1e16-4490-a8a7-b739ea9bc090" (UID: "78b53489-1e16-4490-a8a7-b739ea9bc090"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.233983 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78b53489-1e16-4490-a8a7-b739ea9bc090" (UID: "78b53489-1e16-4490-a8a7-b739ea9bc090"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.236089 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.236120 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfljp\" (UniqueName: \"kubernetes.io/projected/78b53489-1e16-4490-a8a7-b739ea9bc090-kube-api-access-rfljp\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.236132 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78b53489-1e16-4490-a8a7-b739ea9bc090-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.387293 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" path="/var/lib/kubelet/pods/34de0946-abf9-41fd-a0dd-38bf4febbee8/volumes" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.611566 4688 generic.go:334] "Generic (PLEG): container finished" podID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerID="1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527" exitCode=0 Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.611864 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxdmz" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.612361 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerDied","Data":"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527"} Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.612384 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxdmz" event={"ID":"78b53489-1e16-4490-a8a7-b739ea9bc090","Type":"ContainerDied","Data":"582257a138b757787f390a0cc1e2652d602c55340b8955559dcd4505c171e878"} Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.612399 4688 scope.go:117] "RemoveContainer" containerID="1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.624556 4688 scope.go:117] "RemoveContainer" containerID="a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.636647 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.636702 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mxdmz"] Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.645287 4688 scope.go:117] "RemoveContainer" containerID="6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.658807 4688 scope.go:117] "RemoveContainer" containerID="1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527" Oct 01 15:50:13 crc kubenswrapper[4688]: E1001 15:50:13.659990 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527\": container with ID starting with 1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527 not found: ID does not exist" containerID="1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.660033 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527"} err="failed to get container status \"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527\": rpc error: code = NotFound desc = could not find container \"1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527\": container with ID starting with 1a0cbc6c1bf57109922d1c0f5072cb23ceecd41ea38451a2ecb7afcb43b06527 not found: ID does not exist" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.660063 4688 scope.go:117] "RemoveContainer" containerID="a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb" Oct 01 15:50:13 crc kubenswrapper[4688]: E1001 15:50:13.660928 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb\": container with ID starting with a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb not found: ID does not exist" containerID="a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.660961 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb"} err="failed to get container status \"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb\": rpc error: code = NotFound desc = could not find container \"a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb\": container with ID starting with a4b1164061ede616dfe4dc8b0a17287d49f4d6ba5760a08dc9caea2101efc6cb not found: ID does not exist" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.661001 4688 scope.go:117] "RemoveContainer" containerID="6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24" Oct 01 15:50:13 crc kubenswrapper[4688]: E1001 15:50:13.663869 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24\": container with ID starting with 6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24 not found: ID does not exist" containerID="6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24" Oct 01 15:50:13 crc kubenswrapper[4688]: I1001 15:50:13.663899 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24"} err="failed to get container status \"6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24\": rpc error: code = NotFound desc = could not find container \"6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24\": container with ID starting with 6b028d5d12701b4e161e847400403061a36d0de4707cb761e380a76ba6d7bd24 not found: ID does not exist" Oct 01 15:50:15 crc kubenswrapper[4688]: I1001 15:50:15.388410 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" path="/var/lib/kubelet/pods/78b53489-1e16-4490-a8a7-b739ea9bc090/volumes" Oct 01 15:50:16 crc kubenswrapper[4688]: I1001 15:50:16.092601 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:18 crc kubenswrapper[4688]: I1001 15:50:18.955149 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:50:18 crc kubenswrapper[4688]: I1001 15:50:18.955982 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r5sw4" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="registry-server" containerID="cri-o://5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76" gracePeriod=2 Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.332642 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.516657 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities\") pod \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.516758 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4llc\" (UniqueName: \"kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc\") pod \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.516879 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content\") pod \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\" (UID: \"770f9d60-7d86-465b-9ef9-7c0ef8966c5d\") " Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.519216 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities" (OuterVolumeSpecName: "utilities") pod "770f9d60-7d86-465b-9ef9-7c0ef8966c5d" (UID: "770f9d60-7d86-465b-9ef9-7c0ef8966c5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.523477 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc" (OuterVolumeSpecName: "kube-api-access-h4llc") pod "770f9d60-7d86-465b-9ef9-7c0ef8966c5d" (UID: "770f9d60-7d86-465b-9ef9-7c0ef8966c5d"). InnerVolumeSpecName "kube-api-access-h4llc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.582604 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "770f9d60-7d86-465b-9ef9-7c0ef8966c5d" (UID: "770f9d60-7d86-465b-9ef9-7c0ef8966c5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.618935 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.619027 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.619042 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4llc\" (UniqueName: \"kubernetes.io/projected/770f9d60-7d86-465b-9ef9-7c0ef8966c5d-kube-api-access-h4llc\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.645888 4688 generic.go:334] "Generic (PLEG): container finished" podID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerID="5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76" exitCode=0 Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.645925 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerDied","Data":"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76"} Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.645936 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5sw4" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.645951 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5sw4" event={"ID":"770f9d60-7d86-465b-9ef9-7c0ef8966c5d","Type":"ContainerDied","Data":"afeba9525a2ebcc391044283f03b521bce9069221d2547c386caa62d61e1fe03"} Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.645966 4688 scope.go:117] "RemoveContainer" containerID="5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.661719 4688 scope.go:117] "RemoveContainer" containerID="8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.666904 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.682880 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r5sw4"] Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.686697 4688 scope.go:117] "RemoveContainer" containerID="b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.705391 4688 scope.go:117] "RemoveContainer" containerID="5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76" Oct 01 15:50:19 crc kubenswrapper[4688]: E1001 15:50:19.706703 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76\": container with ID starting with 5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76 not found: ID does not exist" containerID="5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.706743 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76"} err="failed to get container status \"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76\": rpc error: code = NotFound desc = could not find container \"5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76\": container with ID starting with 5b0f341d0ef0f90ecaf72deefa441aed9e670fbe64d894d413923a3521334f76 not found: ID does not exist" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.706771 4688 scope.go:117] "RemoveContainer" containerID="8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6" Oct 01 15:50:19 crc kubenswrapper[4688]: E1001 15:50:19.707220 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6\": container with ID starting with 8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6 not found: ID does not exist" containerID="8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.707243 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6"} err="failed to get container status \"8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6\": rpc error: code = NotFound desc = could not find container \"8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6\": container with ID starting with 8f175c64a50d8b6630aa30f271755350f485e07e1bc57bfedc262c6ed79c60b6 not found: ID does not exist" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.707256 4688 scope.go:117] "RemoveContainer" containerID="b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10" Oct 01 15:50:19 crc kubenswrapper[4688]: E1001 15:50:19.707627 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10\": container with ID starting with b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10 not found: ID does not exist" containerID="b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10" Oct 01 15:50:19 crc kubenswrapper[4688]: I1001 15:50:19.707698 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10"} err="failed to get container status \"b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10\": rpc error: code = NotFound desc = could not find container \"b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10\": container with ID starting with b2e553a14324cdc78d99a450e8a4f0004ba80b7a7046a668d1607454cb505a10 not found: ID does not exist" Oct 01 15:50:21 crc kubenswrapper[4688]: I1001 15:50:21.387682 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" path="/var/lib/kubelet/pods/770f9d60-7d86-465b-9ef9-7c0ef8966c5d/volumes" Oct 01 15:50:21 crc kubenswrapper[4688]: I1001 15:50:21.535287 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.052905 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.054627 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.054845 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.055918 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.056150 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5" gracePeriod=600 Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.677067 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5" exitCode=0 Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.677198 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5"} Oct 01 15:50:25 crc kubenswrapper[4688]: I1001 15:50:25.677411 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe"} Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.121553 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" containerID="cri-o://735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1" gracePeriod=15 Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.502684 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.560890 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-86lmd"] Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561275 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561300 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561328 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561337 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561350 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561359 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561377 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e1f5bf-f45c-4801-8b48-ea87e4d57df8" containerName="pruner" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561385 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e1f5bf-f45c-4801-8b48-ea87e4d57df8" containerName="pruner" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561400 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561408 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561419 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561433 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561454 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561462 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561479 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561487 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.561498 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.561506 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.565673 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.565708 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.565721 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.565728 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.565740 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.565746 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.565765 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.565772 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="extract-utilities" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.565781 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.565939 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="extract-content" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567264 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e1f5bf-f45c-4801-8b48-ea87e4d57df8" containerName="pruner" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567286 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b53489-1e16-4490-a8a7-b739ea9bc090" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567318 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2ef8f0-df67-42bf-ad04-5361d0e48c71" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567336 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerName="oauth-openshift" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567354 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f9d60-7d86-465b-9ef9-7c0ef8966c5d" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.567362 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="34de0946-abf9-41fd-a0dd-38bf4febbee8" containerName="registry-server" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.573158 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.575909 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-86lmd"] Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648627 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmqgz\" (UniqueName: \"kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648674 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648712 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648734 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648755 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648770 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648811 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648849 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648875 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648896 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648921 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648949 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648965 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.648995 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig\") pod \"4e28a890-8a72-42ff-9cff-7c3f100abdce\" (UID: \"4e28a890-8a72-42ff-9cff-7c3f100abdce\") " Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.649778 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.650053 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.650185 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.650679 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.651152 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.654243 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.654549 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.655980 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz" (OuterVolumeSpecName: "kube-api-access-pmqgz") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "kube-api-access-pmqgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.655874 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.657726 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.657870 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.658237 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.658746 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.658960 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4e28a890-8a72-42ff-9cff-7c3f100abdce" (UID: "4e28a890-8a72-42ff-9cff-7c3f100abdce"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.749913 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.749970 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.749996 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750035 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750058 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-dir\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750091 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750130 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750169 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750249 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750272 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-policies\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750307 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750332 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750516 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750713 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9sx4\" (UniqueName: \"kubernetes.io/projected/b5cb44d1-ca80-4328-87bb-90afe62534e2-kube-api-access-c9sx4\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750809 4688 generic.go:334] "Generic (PLEG): container finished" podID="4e28a890-8a72-42ff-9cff-7c3f100abdce" containerID="735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1" exitCode=0 Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750855 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" event={"ID":"4e28a890-8a72-42ff-9cff-7c3f100abdce","Type":"ContainerDied","Data":"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1"} Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750888 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750903 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m94rs" event={"ID":"4e28a890-8a72-42ff-9cff-7c3f100abdce","Type":"ContainerDied","Data":"18a0351a0b623fb86ab1c2f1692dde410ee264794d57e92805969dc2723bfb07"} Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750905 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750923 4688 scope.go:117] "RemoveContainer" containerID="735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750949 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750973 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.750993 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751013 4688 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751031 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751102 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751452 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751476 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751582 4688 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4e28a890-8a72-42ff-9cff-7c3f100abdce-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751603 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751624 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751641 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmqgz\" (UniqueName: \"kubernetes.io/projected/4e28a890-8a72-42ff-9cff-7c3f100abdce-kube-api-access-pmqgz\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.751658 4688 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4e28a890-8a72-42ff-9cff-7c3f100abdce-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.770850 4688 scope.go:117] "RemoveContainer" containerID="735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1" Oct 01 15:50:37 crc kubenswrapper[4688]: E1001 15:50:37.771366 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1\": container with ID starting with 735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1 not found: ID does not exist" containerID="735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.771416 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1"} err="failed to get container status \"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1\": rpc error: code = NotFound desc = could not find container \"735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1\": container with ID starting with 735a4160b488a8ca2ae759c3709ae08557f41edff891d25a24c5c2f5a650c5c1 not found: ID does not exist" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.791927 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.795256 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m94rs"] Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852416 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852469 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852506 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852584 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852619 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-policies\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852651 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852684 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852718 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852750 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9sx4\" (UniqueName: \"kubernetes.io/projected/b5cb44d1-ca80-4328-87bb-90afe62534e2-kube-api-access-c9sx4\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852809 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852843 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852871 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852914 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.852947 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-dir\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.853071 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-dir\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.854989 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.856284 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.856446 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-audit-policies\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.858578 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.861231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.862319 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.871218 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.871308 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.871603 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.871883 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.872103 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.881668 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5cb44d1-ca80-4328-87bb-90afe62534e2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.884901 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9sx4\" (UniqueName: \"kubernetes.io/projected/b5cb44d1-ca80-4328-87bb-90afe62534e2-kube-api-access-c9sx4\") pod \"oauth-openshift-657494565c-86lmd\" (UID: \"b5cb44d1-ca80-4328-87bb-90afe62534e2\") " pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:37 crc kubenswrapper[4688]: I1001 15:50:37.898230 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.344239 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-86lmd"] Oct 01 15:50:38 crc kubenswrapper[4688]: W1001 15:50:38.350713 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5cb44d1_ca80_4328_87bb_90afe62534e2.slice/crio-98f0bd63c63855c9e35e2d4d35b4edb85cd6e65f2bdfb051b6676c0495d1786e WatchSource:0}: Error finding container 98f0bd63c63855c9e35e2d4d35b4edb85cd6e65f2bdfb051b6676c0495d1786e: Status 404 returned error can't find the container with id 98f0bd63c63855c9e35e2d4d35b4edb85cd6e65f2bdfb051b6676c0495d1786e Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.758266 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" event={"ID":"b5cb44d1-ca80-4328-87bb-90afe62534e2","Type":"ContainerStarted","Data":"b39b26713b1b7cf3fc7305ab86d9ad2e50f90de96e02c679e716ef18fbdf7174"} Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.758325 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" event={"ID":"b5cb44d1-ca80-4328-87bb-90afe62534e2","Type":"ContainerStarted","Data":"98f0bd63c63855c9e35e2d4d35b4edb85cd6e65f2bdfb051b6676c0495d1786e"} Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.759196 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.799293 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" podStartSLOduration=26.799268539 podStartE2EDuration="26.799268539s" podCreationTimestamp="2025-10-01 15:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:50:38.796707282 +0000 UTC m=+228.147347294" watchObservedRunningTime="2025-10-01 15:50:38.799268539 +0000 UTC m=+228.149908531" Oct 01 15:50:38 crc kubenswrapper[4688]: I1001 15:50:38.981119 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-657494565c-86lmd" Oct 01 15:50:39 crc kubenswrapper[4688]: I1001 15:50:39.391329 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e28a890-8a72-42ff-9cff-7c3f100abdce" path="/var/lib/kubelet/pods/4e28a890-8a72-42ff-9cff-7c3f100abdce/volumes" Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.859050 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.859890 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gkhsz" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="registry-server" containerID="cri-o://badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164" gracePeriod=30 Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.864980 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.865193 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vj4w9" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="registry-server" containerID="cri-o://4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3" gracePeriod=30 Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.887398 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.887992 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" containerID="cri-o://2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522" gracePeriod=30 Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.907047 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.907408 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kp29h" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="registry-server" containerID="cri-o://f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2" gracePeriod=30 Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.921262 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.921830 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kcmw2" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="registry-server" containerID="cri-o://086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af" gracePeriod=30 Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.925684 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kszs7"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.927341 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.932686 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kszs7"] Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.936347 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.936421 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:51 crc kubenswrapper[4688]: I1001 15:50:51.936462 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkvxr\" (UniqueName: \"kubernetes.io/projected/a368e405-b690-47e8-b2e7-ff9917147420-kube-api-access-gkvxr\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.042547 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.042619 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.042673 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkvxr\" (UniqueName: \"kubernetes.io/projected/a368e405-b690-47e8-b2e7-ff9917147420-kube-api-access-gkvxr\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.045507 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.053408 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a368e405-b690-47e8-b2e7-ff9917147420-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.068499 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkvxr\" (UniqueName: \"kubernetes.io/projected/a368e405-b690-47e8-b2e7-ff9917147420-kube-api-access-gkvxr\") pod \"marketplace-operator-79b997595-kszs7\" (UID: \"a368e405-b690-47e8-b2e7-ff9917147420\") " pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.252816 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.269021 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.347665 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities\") pod \"ad746ee8-2f78-40b9-b980-286172f023d4\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.348139 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content\") pod \"ad746ee8-2f78-40b9-b980-286172f023d4\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.348248 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbxf4\" (UniqueName: \"kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4\") pod \"ad746ee8-2f78-40b9-b980-286172f023d4\" (UID: \"ad746ee8-2f78-40b9-b980-286172f023d4\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.349124 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities" (OuterVolumeSpecName: "utilities") pod "ad746ee8-2f78-40b9-b980-286172f023d4" (UID: "ad746ee8-2f78-40b9-b980-286172f023d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.357682 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4" (OuterVolumeSpecName: "kube-api-access-dbxf4") pod "ad746ee8-2f78-40b9-b980-286172f023d4" (UID: "ad746ee8-2f78-40b9-b980-286172f023d4"). InnerVolumeSpecName "kube-api-access-dbxf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.360710 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.392364 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.418671 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.436137 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450164 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7kcj\" (UniqueName: \"kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj\") pod \"d54684b4-efba-4ebf-91bf-524247aafc4e\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450209 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sd8h\" (UniqueName: \"kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h\") pod \"7c051410-74cd-4aad-9090-437673913d17\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450242 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities\") pod \"3f036344-df58-4ea8-b960-28908e1cac82\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450299 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics\") pod \"7c051410-74cd-4aad-9090-437673913d17\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450330 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content\") pod \"3f036344-df58-4ea8-b960-28908e1cac82\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450353 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities\") pod \"d54684b4-efba-4ebf-91bf-524247aafc4e\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450378 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk9rl\" (UniqueName: \"kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl\") pod \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450706 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.450724 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbxf4\" (UniqueName: \"kubernetes.io/projected/ad746ee8-2f78-40b9-b980-286172f023d4-kube-api-access-dbxf4\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.452068 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities" (OuterVolumeSpecName: "utilities") pod "3f036344-df58-4ea8-b960-28908e1cac82" (UID: "3f036344-df58-4ea8-b960-28908e1cac82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.454519 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad746ee8-2f78-40b9-b980-286172f023d4" (UID: "ad746ee8-2f78-40b9-b980-286172f023d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.455173 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities" (OuterVolumeSpecName: "utilities") pod "d54684b4-efba-4ebf-91bf-524247aafc4e" (UID: "d54684b4-efba-4ebf-91bf-524247aafc4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.461218 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj" (OuterVolumeSpecName: "kube-api-access-p7kcj") pod "d54684b4-efba-4ebf-91bf-524247aafc4e" (UID: "d54684b4-efba-4ebf-91bf-524247aafc4e"). InnerVolumeSpecName "kube-api-access-p7kcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.462970 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7c051410-74cd-4aad-9090-437673913d17" (UID: "7c051410-74cd-4aad-9090-437673913d17"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.468132 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl" (OuterVolumeSpecName: "kube-api-access-pk9rl") pod "e1cb0316-bcd2-46ad-8f23-db3c545eff1c" (UID: "e1cb0316-bcd2-46ad-8f23-db3c545eff1c"). InnerVolumeSpecName "kube-api-access-pk9rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.473926 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f036344-df58-4ea8-b960-28908e1cac82" (UID: "3f036344-df58-4ea8-b960-28908e1cac82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.477391 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h" (OuterVolumeSpecName: "kube-api-access-6sd8h") pod "7c051410-74cd-4aad-9090-437673913d17" (UID: "7c051410-74cd-4aad-9090-437673913d17"). InnerVolumeSpecName "kube-api-access-6sd8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.551572 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities\") pod \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.551632 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtgrx\" (UniqueName: \"kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx\") pod \"3f036344-df58-4ea8-b960-28908e1cac82\" (UID: \"3f036344-df58-4ea8-b960-28908e1cac82\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.551674 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content\") pod \"d54684b4-efba-4ebf-91bf-524247aafc4e\" (UID: \"d54684b4-efba-4ebf-91bf-524247aafc4e\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.551724 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca\") pod \"7c051410-74cd-4aad-9090-437673913d17\" (UID: \"7c051410-74cd-4aad-9090-437673913d17\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.551749 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content\") pod \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\" (UID: \"e1cb0316-bcd2-46ad-8f23-db3c545eff1c\") " Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552028 4688 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7c051410-74cd-4aad-9090-437673913d17-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552040 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552049 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552057 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk9rl\" (UniqueName: \"kubernetes.io/projected/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-kube-api-access-pk9rl\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552066 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad746ee8-2f78-40b9-b980-286172f023d4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552075 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7kcj\" (UniqueName: \"kubernetes.io/projected/d54684b4-efba-4ebf-91bf-524247aafc4e-kube-api-access-p7kcj\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552083 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sd8h\" (UniqueName: \"kubernetes.io/projected/7c051410-74cd-4aad-9090-437673913d17-kube-api-access-6sd8h\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552091 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f036344-df58-4ea8-b960-28908e1cac82-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552308 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities" (OuterVolumeSpecName: "utilities") pod "e1cb0316-bcd2-46ad-8f23-db3c545eff1c" (UID: "e1cb0316-bcd2-46ad-8f23-db3c545eff1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.552905 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7c051410-74cd-4aad-9090-437673913d17" (UID: "7c051410-74cd-4aad-9090-437673913d17"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.554453 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx" (OuterVolumeSpecName: "kube-api-access-xtgrx") pod "3f036344-df58-4ea8-b960-28908e1cac82" (UID: "3f036344-df58-4ea8-b960-28908e1cac82"). InnerVolumeSpecName "kube-api-access-xtgrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.605579 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1cb0316-bcd2-46ad-8f23-db3c545eff1c" (UID: "e1cb0316-bcd2-46ad-8f23-db3c545eff1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.649163 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d54684b4-efba-4ebf-91bf-524247aafc4e" (UID: "d54684b4-efba-4ebf-91bf-524247aafc4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.653390 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.653426 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtgrx\" (UniqueName: \"kubernetes.io/projected/3f036344-df58-4ea8-b960-28908e1cac82-kube-api-access-xtgrx\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.653437 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d54684b4-efba-4ebf-91bf-524247aafc4e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.653450 4688 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c051410-74cd-4aad-9090-437673913d17-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.653459 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb0316-bcd2-46ad-8f23-db3c545eff1c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.818401 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kszs7"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.854696 4688 generic.go:334] "Generic (PLEG): container finished" podID="3f036344-df58-4ea8-b960-28908e1cac82" containerID="f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2" exitCode=0 Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.854758 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp29h" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.855038 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerDied","Data":"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.855145 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp29h" event={"ID":"3f036344-df58-4ea8-b960-28908e1cac82","Type":"ContainerDied","Data":"f478b7294a733e4b61a3258c014d10dfdc7a56bcf4ab8be501421df944aa2ded"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.855180 4688 scope.go:117] "RemoveContainer" containerID="f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.859169 4688 generic.go:334] "Generic (PLEG): container finished" podID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerID="badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164" exitCode=0 Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.859222 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerDied","Data":"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.859245 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkhsz" event={"ID":"e1cb0316-bcd2-46ad-8f23-db3c545eff1c","Type":"ContainerDied","Data":"b6d698527955d0614613f9f54507065c03f6b5be3d89365f9851150983e74bb7"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.859297 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkhsz" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.867020 4688 generic.go:334] "Generic (PLEG): container finished" podID="7c051410-74cd-4aad-9090-437673913d17" containerID="2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522" exitCode=0 Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.867182 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.867587 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" event={"ID":"7c051410-74cd-4aad-9090-437673913d17","Type":"ContainerDied","Data":"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.867643 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vfflp" event={"ID":"7c051410-74cd-4aad-9090-437673913d17","Type":"ContainerDied","Data":"a9ee81390f04f50caebf1c18c98160ac04d0017503642ce3259ea20da5cfa8c6"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.874098 4688 generic.go:334] "Generic (PLEG): container finished" podID="ad746ee8-2f78-40b9-b980-286172f023d4" containerID="4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3" exitCode=0 Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.874246 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj4w9" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.874656 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerDied","Data":"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.874693 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj4w9" event={"ID":"ad746ee8-2f78-40b9-b980-286172f023d4","Type":"ContainerDied","Data":"17e03b5148bb12ceeee5e14b32b5af0863c2fd5bfb98ca173e4b5684223e8ba7"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.875985 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" event={"ID":"a368e405-b690-47e8-b2e7-ff9917147420","Type":"ContainerStarted","Data":"dccfdabc5b25306e0ea4d4d434b01e9ce618601625ef05386ddb0de9b7d380e7"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.882127 4688 generic.go:334] "Generic (PLEG): container finished" podID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerID="086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af" exitCode=0 Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.882164 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerDied","Data":"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.882189 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcmw2" event={"ID":"d54684b4-efba-4ebf-91bf-524247aafc4e","Type":"ContainerDied","Data":"b27f3491a54d0d0c59e6a5d84c6f95f3920682698412b85cefaa07164a0d183a"} Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.882276 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcmw2" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.893459 4688 scope.go:117] "RemoveContainer" containerID="23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.914199 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.921757 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp29h"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.936546 4688 scope.go:117] "RemoveContainer" containerID="dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.941051 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.952037 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gkhsz"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.959570 4688 scope.go:117] "RemoveContainer" containerID="f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.959709 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:50:52 crc kubenswrapper[4688]: E1001 15:50:52.966035 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2\": container with ID starting with f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2 not found: ID does not exist" containerID="f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.966081 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2"} err="failed to get container status \"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2\": rpc error: code = NotFound desc = could not find container \"f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2\": container with ID starting with f9ea0c853343fe6190cafddb57b31391a4020f138e19d8a262e8bb3300b99ae2 not found: ID does not exist" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.966113 4688 scope.go:117] "RemoveContainer" containerID="23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe" Oct 01 15:50:52 crc kubenswrapper[4688]: E1001 15:50:52.967631 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe\": container with ID starting with 23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe not found: ID does not exist" containerID="23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.967664 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe"} err="failed to get container status \"23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe\": rpc error: code = NotFound desc = could not find container \"23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe\": container with ID starting with 23367c7dae34071f522c19ca44892bd12a735010e181d9b0ca9ddca0dc615abe not found: ID does not exist" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.967684 4688 scope.go:117] "RemoveContainer" containerID="dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.969146 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vfflp"] Oct 01 15:50:52 crc kubenswrapper[4688]: E1001 15:50:52.970992 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088\": container with ID starting with dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088 not found: ID does not exist" containerID="dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.971026 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088"} err="failed to get container status \"dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088\": rpc error: code = NotFound desc = could not find container \"dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088\": container with ID starting with dfd4972deb677f701096079f8bde0582202cea17617bfd76e1b7d2e876be0088 not found: ID does not exist" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.971046 4688 scope.go:117] "RemoveContainer" containerID="badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164" Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.972268 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.974591 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vj4w9"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.977173 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.979232 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kcmw2"] Oct 01 15:50:52 crc kubenswrapper[4688]: I1001 15:50:52.988608 4688 scope.go:117] "RemoveContainer" containerID="ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.039760 4688 scope.go:117] "RemoveContainer" containerID="382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.051232 4688 scope.go:117] "RemoveContainer" containerID="badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.054790 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164\": container with ID starting with badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164 not found: ID does not exist" containerID="badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.054912 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164"} err="failed to get container status \"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164\": rpc error: code = NotFound desc = could not find container \"badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164\": container with ID starting with badb4f7126169f4c3e8ee20a242c10d6be2c97889de336c34e5ef6ef69eb5164 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.055014 4688 scope.go:117] "RemoveContainer" containerID="ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.057025 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a\": container with ID starting with ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a not found: ID does not exist" containerID="ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.057128 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a"} err="failed to get container status \"ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a\": rpc error: code = NotFound desc = could not find container \"ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a\": container with ID starting with ab284095df3f5f9a57443277a408e46df54e21316cf767e223246efa9db7f30a not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.057207 4688 scope.go:117] "RemoveContainer" containerID="382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.057469 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629\": container with ID starting with 382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629 not found: ID does not exist" containerID="382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.057556 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629"} err="failed to get container status \"382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629\": rpc error: code = NotFound desc = could not find container \"382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629\": container with ID starting with 382bb0303d94fdb0ff1f93f0f5d9aa50a7c5230af234101087ce968b9f7c6629 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.057632 4688 scope.go:117] "RemoveContainer" containerID="2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.068177 4688 scope.go:117] "RemoveContainer" containerID="2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.068679 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522\": container with ID starting with 2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522 not found: ID does not exist" containerID="2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.068781 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522"} err="failed to get container status \"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522\": rpc error: code = NotFound desc = could not find container \"2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522\": container with ID starting with 2047aa9aacda862e5a43da45f7005508cbaf21b604141e8bbc64e2fe45ab9522 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.068876 4688 scope.go:117] "RemoveContainer" containerID="4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.081880 4688 scope.go:117] "RemoveContainer" containerID="8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.097432 4688 scope.go:117] "RemoveContainer" containerID="eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.107204 4688 scope.go:117] "RemoveContainer" containerID="4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.107456 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3\": container with ID starting with 4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3 not found: ID does not exist" containerID="4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.107484 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3"} err="failed to get container status \"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3\": rpc error: code = NotFound desc = could not find container \"4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3\": container with ID starting with 4de1b9bf4518730a82d6c8983f9fcfe66713fb01118e7c8f19877578219c81d3 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.107504 4688 scope.go:117] "RemoveContainer" containerID="8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.107859 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7\": container with ID starting with 8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7 not found: ID does not exist" containerID="8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.107884 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7"} err="failed to get container status \"8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7\": rpc error: code = NotFound desc = could not find container \"8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7\": container with ID starting with 8d5163f0170640f9df96034e1557d0b8219b2de8385e2437b83dd84a185c23d7 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.107899 4688 scope.go:117] "RemoveContainer" containerID="eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.108201 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c\": container with ID starting with eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c not found: ID does not exist" containerID="eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.108221 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c"} err="failed to get container status \"eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c\": rpc error: code = NotFound desc = could not find container \"eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c\": container with ID starting with eb27402addb4291710ccb8a2367aa52b570879868f0f6ba7f4624eeca23c899c not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.108233 4688 scope.go:117] "RemoveContainer" containerID="086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.125250 4688 scope.go:117] "RemoveContainer" containerID="1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.136584 4688 scope.go:117] "RemoveContainer" containerID="70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.148441 4688 scope.go:117] "RemoveContainer" containerID="086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.148710 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af\": container with ID starting with 086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af not found: ID does not exist" containerID="086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.148740 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af"} err="failed to get container status \"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af\": rpc error: code = NotFound desc = could not find container \"086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af\": container with ID starting with 086331e54dfb672c25cbf822c338d0d71d25e4a9b8c5f795e3ef92ba997d93af not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.148770 4688 scope.go:117] "RemoveContainer" containerID="1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.149316 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad\": container with ID starting with 1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad not found: ID does not exist" containerID="1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.149343 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad"} err="failed to get container status \"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad\": rpc error: code = NotFound desc = could not find container \"1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad\": container with ID starting with 1fa86e540e8a0b5d0dd44433298edf9ba9b4dc2508334a11b9169ec2282ce0ad not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.149359 4688 scope.go:117] "RemoveContainer" containerID="70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3" Oct 01 15:50:53 crc kubenswrapper[4688]: E1001 15:50:53.149605 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3\": container with ID starting with 70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3 not found: ID does not exist" containerID="70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.149633 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3"} err="failed to get container status \"70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3\": rpc error: code = NotFound desc = could not find container \"70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3\": container with ID starting with 70665a80899ca69728e23d87a07eff4b9f1edb15f72570d0c10e1ee129b413c3 not found: ID does not exist" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.401226 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f036344-df58-4ea8-b960-28908e1cac82" path="/var/lib/kubelet/pods/3f036344-df58-4ea8-b960-28908e1cac82/volumes" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.402005 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c051410-74cd-4aad-9090-437673913d17" path="/var/lib/kubelet/pods/7c051410-74cd-4aad-9090-437673913d17/volumes" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.402546 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" path="/var/lib/kubelet/pods/ad746ee8-2f78-40b9-b980-286172f023d4/volumes" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.403736 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" path="/var/lib/kubelet/pods/d54684b4-efba-4ebf-91bf-524247aafc4e/volumes" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.404387 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" path="/var/lib/kubelet/pods/e1cb0316-bcd2-46ad-8f23-db3c545eff1c/volumes" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.906561 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" event={"ID":"a368e405-b690-47e8-b2e7-ff9917147420","Type":"ContainerStarted","Data":"163a6a570c1352e1e9ebb276bd26e8f5d307c7fec73b2057b5483aa2b1d4426c"} Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.909707 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.920042 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" Oct 01 15:50:53 crc kubenswrapper[4688]: I1001 15:50:53.946845 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kszs7" podStartSLOduration=2.946825199 podStartE2EDuration="2.946825199s" podCreationTimestamp="2025-10-01 15:50:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:50:53.924572623 +0000 UTC m=+243.275212585" watchObservedRunningTime="2025-10-01 15:50:53.946825199 +0000 UTC m=+243.297465161" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.083350 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m242n"] Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.085899 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.085924 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.085937 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.085945 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.085955 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.085970 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.085983 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.085997 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086007 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086015 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086024 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086031 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086042 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086049 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086060 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086068 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086086 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086094 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086104 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086111 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086123 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086130 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086138 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086145 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="extract-content" Oct 01 15:50:54 crc kubenswrapper[4688]: E1001 15:50:54.086154 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086162 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="extract-utilities" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086275 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d54684b4-efba-4ebf-91bf-524247aafc4e" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086286 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad746ee8-2f78-40b9-b980-286172f023d4" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086297 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f036344-df58-4ea8-b960-28908e1cac82" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086308 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1cb0316-bcd2-46ad-8f23-db3c545eff1c" containerName="registry-server" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.086323 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c051410-74cd-4aad-9090-437673913d17" containerName="marketplace-operator" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.087121 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.091729 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.092684 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m242n"] Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.169436 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-catalog-content\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.169783 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-utilities\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.169870 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4f6v\" (UniqueName: \"kubernetes.io/projected/092133e9-cde7-4be0-b5c2-e5de34a02727-kube-api-access-t4f6v\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.270796 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-utilities\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.270852 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4f6v\" (UniqueName: \"kubernetes.io/projected/092133e9-cde7-4be0-b5c2-e5de34a02727-kube-api-access-t4f6v\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.270885 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-catalog-content\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.271333 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-catalog-content\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.271514 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092133e9-cde7-4be0-b5c2-e5de34a02727-utilities\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.282960 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.285458 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.287602 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.307723 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4f6v\" (UniqueName: \"kubernetes.io/projected/092133e9-cde7-4be0-b5c2-e5de34a02727-kube-api-access-t4f6v\") pod \"redhat-marketplace-m242n\" (UID: \"092133e9-cde7-4be0-b5c2-e5de34a02727\") " pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.308699 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.405991 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.472628 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.473254 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.473696 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kstp9\" (UniqueName: \"kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.574581 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kstp9\" (UniqueName: \"kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.574639 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.574679 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.575292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.575337 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.594348 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kstp9\" (UniqueName: \"kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9\") pod \"redhat-operators-wq6zm\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.625216 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.786870 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m242n"] Oct 01 15:50:54 crc kubenswrapper[4688]: W1001 15:50:54.795632 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod092133e9_cde7_4be0_b5c2_e5de34a02727.slice/crio-5d4b7ef9391982150afea8d46e2e966c49c8fb0b7b37355357fcc8b3bb5784f9 WatchSource:0}: Error finding container 5d4b7ef9391982150afea8d46e2e966c49c8fb0b7b37355357fcc8b3bb5784f9: Status 404 returned error can't find the container with id 5d4b7ef9391982150afea8d46e2e966c49c8fb0b7b37355357fcc8b3bb5784f9 Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.930183 4688 generic.go:334] "Generic (PLEG): container finished" podID="092133e9-cde7-4be0-b5c2-e5de34a02727" containerID="6398ab83a440251bff5488fb4ebf5e1e14796b605af7c939484157c8115d9a26" exitCode=0 Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.931204 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m242n" event={"ID":"092133e9-cde7-4be0-b5c2-e5de34a02727","Type":"ContainerDied","Data":"6398ab83a440251bff5488fb4ebf5e1e14796b605af7c939484157c8115d9a26"} Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.931228 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m242n" event={"ID":"092133e9-cde7-4be0-b5c2-e5de34a02727","Type":"ContainerStarted","Data":"5d4b7ef9391982150afea8d46e2e966c49c8fb0b7b37355357fcc8b3bb5784f9"} Oct 01 15:50:54 crc kubenswrapper[4688]: I1001 15:50:54.986853 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 15:50:54 crc kubenswrapper[4688]: W1001 15:50:54.991715 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd15a0ff9_a537_4e8b_8836_5b20538fe681.slice/crio-d61ec1d1903814d8845bdec47079d03a1b948acc4f25f6c50aac80c35312d9ee WatchSource:0}: Error finding container d61ec1d1903814d8845bdec47079d03a1b948acc4f25f6c50aac80c35312d9ee: Status 404 returned error can't find the container with id d61ec1d1903814d8845bdec47079d03a1b948acc4f25f6c50aac80c35312d9ee Oct 01 15:50:55 crc kubenswrapper[4688]: I1001 15:50:55.948118 4688 generic.go:334] "Generic (PLEG): container finished" podID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerID="9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401" exitCode=0 Oct 01 15:50:55 crc kubenswrapper[4688]: I1001 15:50:55.949379 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerDied","Data":"9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401"} Oct 01 15:50:55 crc kubenswrapper[4688]: I1001 15:50:55.949410 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerStarted","Data":"d61ec1d1903814d8845bdec47079d03a1b948acc4f25f6c50aac80c35312d9ee"} Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.486654 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gpqvx"] Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.487817 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.490840 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.493586 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-catalog-content\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.493650 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmbkb\" (UniqueName: \"kubernetes.io/projected/5df443ab-56dd-4124-963e-8e109914ea7e-kube-api-access-fmbkb\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.493729 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-utilities\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.510076 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpqvx"] Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.594221 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-utilities\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.594392 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-catalog-content\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.594435 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmbkb\" (UniqueName: \"kubernetes.io/projected/5df443ab-56dd-4124-963e-8e109914ea7e-kube-api-access-fmbkb\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.594717 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-utilities\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.594796 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5df443ab-56dd-4124-963e-8e109914ea7e-catalog-content\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.620205 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmbkb\" (UniqueName: \"kubernetes.io/projected/5df443ab-56dd-4124-963e-8e109914ea7e-kube-api-access-fmbkb\") pod \"community-operators-gpqvx\" (UID: \"5df443ab-56dd-4124-963e-8e109914ea7e\") " pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.682123 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.688204 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.690696 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.692871 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.695561 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ngmc\" (UniqueName: \"kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.695698 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.695820 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.796542 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ngmc\" (UniqueName: \"kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.796587 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.796622 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.797001 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.797070 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.802687 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.813944 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ngmc\" (UniqueName: \"kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc\") pod \"certified-operators-w5tln\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.965196 4688 generic.go:334] "Generic (PLEG): container finished" podID="092133e9-cde7-4be0-b5c2-e5de34a02727" containerID="b0d0bfa632068e728cd95da3f1b0edc1052cdb0c03be8af86723356881927626" exitCode=0 Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.965502 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m242n" event={"ID":"092133e9-cde7-4be0-b5c2-e5de34a02727","Type":"ContainerDied","Data":"b0d0bfa632068e728cd95da3f1b0edc1052cdb0c03be8af86723356881927626"} Oct 01 15:50:56 crc kubenswrapper[4688]: I1001 15:50:56.975577 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerStarted","Data":"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.003776 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.223590 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpqvx"] Oct 01 15:50:57 crc kubenswrapper[4688]: W1001 15:50:57.228771 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5df443ab_56dd_4124_963e_8e109914ea7e.slice/crio-f50349f16de824a1d3c4d7acb49313743f9fbac7942497e04c15883b6c6aa0f2 WatchSource:0}: Error finding container f50349f16de824a1d3c4d7acb49313743f9fbac7942497e04c15883b6c6aa0f2: Status 404 returned error can't find the container with id f50349f16de824a1d3c4d7acb49313743f9fbac7942497e04c15883b6c6aa0f2 Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.403497 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.985436 4688 generic.go:334] "Generic (PLEG): container finished" podID="245214e6-aa79-4963-9524-2c6852196d70" containerID="aa5178bc6063b3df9716bd3f7f2317daaf045c2a33f5d6c49c3b2d8988105f5e" exitCode=0 Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.985473 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerDied","Data":"aa5178bc6063b3df9716bd3f7f2317daaf045c2a33f5d6c49c3b2d8988105f5e"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.985896 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerStarted","Data":"ad3d10906e73b8eec77cec519812024817370fc9bcd9a2efb5638289b8c1ac96"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.988650 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m242n" event={"ID":"092133e9-cde7-4be0-b5c2-e5de34a02727","Type":"ContainerStarted","Data":"e4eae1ccd5fcd0f83aeb9c83f267b424cf1b14608d6b6812885c0611c57cb814"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.990046 4688 generic.go:334] "Generic (PLEG): container finished" podID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerID="15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f" exitCode=0 Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.990085 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerDied","Data":"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.995909 4688 generic.go:334] "Generic (PLEG): container finished" podID="5df443ab-56dd-4124-963e-8e109914ea7e" containerID="535e4128de9d5194267bacc1ee8132c7852f48e200436c74905b84294d6594f1" exitCode=0 Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.995943 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpqvx" event={"ID":"5df443ab-56dd-4124-963e-8e109914ea7e","Type":"ContainerDied","Data":"535e4128de9d5194267bacc1ee8132c7852f48e200436c74905b84294d6594f1"} Oct 01 15:50:57 crc kubenswrapper[4688]: I1001 15:50:57.995966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpqvx" event={"ID":"5df443ab-56dd-4124-963e-8e109914ea7e","Type":"ContainerStarted","Data":"f50349f16de824a1d3c4d7acb49313743f9fbac7942497e04c15883b6c6aa0f2"} Oct 01 15:50:58 crc kubenswrapper[4688]: I1001 15:50:58.067318 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m242n" podStartSLOduration=1.503138065 podStartE2EDuration="4.067300133s" podCreationTimestamp="2025-10-01 15:50:54 +0000 UTC" firstStartedPulling="2025-10-01 15:50:54.932343582 +0000 UTC m=+244.282983544" lastFinishedPulling="2025-10-01 15:50:57.49650565 +0000 UTC m=+246.847145612" observedRunningTime="2025-10-01 15:50:58.065256243 +0000 UTC m=+247.415896225" watchObservedRunningTime="2025-10-01 15:50:58.067300133 +0000 UTC m=+247.417940095" Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.020237 4688 generic.go:334] "Generic (PLEG): container finished" podID="5df443ab-56dd-4124-963e-8e109914ea7e" containerID="d81825c9114404b498dda631d11b527861e19bd3ae9e6ca36586adc01feda298" exitCode=0 Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.020711 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpqvx" event={"ID":"5df443ab-56dd-4124-963e-8e109914ea7e","Type":"ContainerDied","Data":"d81825c9114404b498dda631d11b527861e19bd3ae9e6ca36586adc01feda298"} Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.023695 4688 generic.go:334] "Generic (PLEG): container finished" podID="245214e6-aa79-4963-9524-2c6852196d70" containerID="59363691d0fbd2163afb8331702d6ec613069b8716ffc4ea38af62a20a7a11ff" exitCode=0 Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.023761 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerDied","Data":"59363691d0fbd2163afb8331702d6ec613069b8716ffc4ea38af62a20a7a11ff"} Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.031309 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerStarted","Data":"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa"} Oct 01 15:51:01 crc kubenswrapper[4688]: I1001 15:51:01.090399 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wq6zm" podStartSLOduration=3.984407806 podStartE2EDuration="7.090379132s" podCreationTimestamp="2025-10-01 15:50:54 +0000 UTC" firstStartedPulling="2025-10-01 15:50:55.949932335 +0000 UTC m=+245.300572287" lastFinishedPulling="2025-10-01 15:50:59.055903611 +0000 UTC m=+248.406543613" observedRunningTime="2025-10-01 15:51:01.087708972 +0000 UTC m=+250.438348974" watchObservedRunningTime="2025-10-01 15:51:01.090379132 +0000 UTC m=+250.441019084" Oct 01 15:51:02 crc kubenswrapper[4688]: I1001 15:51:02.037803 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpqvx" event={"ID":"5df443ab-56dd-4124-963e-8e109914ea7e","Type":"ContainerStarted","Data":"35d6a30b44bb2439dfbc711f22bc2a053c563815240dc60325e68f343dc5a0cb"} Oct 01 15:51:02 crc kubenswrapper[4688]: I1001 15:51:02.040385 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerStarted","Data":"9baddb6f9d1506db1adbec28541f0d94ea4a8d2e5d1e7b284a0f804dc7d26402"} Oct 01 15:51:02 crc kubenswrapper[4688]: I1001 15:51:02.078394 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w5tln" podStartSLOduration=2.420081799 podStartE2EDuration="6.07837483s" podCreationTimestamp="2025-10-01 15:50:56 +0000 UTC" firstStartedPulling="2025-10-01 15:50:57.986688014 +0000 UTC m=+247.337327976" lastFinishedPulling="2025-10-01 15:51:01.644981045 +0000 UTC m=+250.995621007" observedRunningTime="2025-10-01 15:51:02.07750089 +0000 UTC m=+251.428140852" watchObservedRunningTime="2025-10-01 15:51:02.07837483 +0000 UTC m=+251.429014792" Oct 01 15:51:02 crc kubenswrapper[4688]: I1001 15:51:02.081298 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gpqvx" podStartSLOduration=2.627303359 podStartE2EDuration="6.081289789s" podCreationTimestamp="2025-10-01 15:50:56 +0000 UTC" firstStartedPulling="2025-10-01 15:50:57.996989644 +0000 UTC m=+247.347629606" lastFinishedPulling="2025-10-01 15:51:01.450976074 +0000 UTC m=+250.801616036" observedRunningTime="2025-10-01 15:51:02.063198134 +0000 UTC m=+251.413838096" watchObservedRunningTime="2025-10-01 15:51:02.081289789 +0000 UTC m=+251.431929751" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.406358 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.406756 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.454782 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.626289 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.626349 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:51:04 crc kubenswrapper[4688]: I1001 15:51:04.663559 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:51:05 crc kubenswrapper[4688]: I1001 15:51:05.095203 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 15:51:05 crc kubenswrapper[4688]: I1001 15:51:05.100800 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m242n" Oct 01 15:51:06 crc kubenswrapper[4688]: I1001 15:51:06.803613 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:51:06 crc kubenswrapper[4688]: I1001 15:51:06.803669 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:51:06 crc kubenswrapper[4688]: I1001 15:51:06.854457 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:51:07 crc kubenswrapper[4688]: I1001 15:51:07.004773 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:51:07 crc kubenswrapper[4688]: I1001 15:51:07.005006 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:51:07 crc kubenswrapper[4688]: I1001 15:51:07.064751 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:51:07 crc kubenswrapper[4688]: I1001 15:51:07.108169 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 15:51:07 crc kubenswrapper[4688]: I1001 15:51:07.108239 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gpqvx" Oct 01 15:52:25 crc kubenswrapper[4688]: I1001 15:52:25.053046 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:52:25 crc kubenswrapper[4688]: I1001 15:52:25.054797 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:52:55 crc kubenswrapper[4688]: I1001 15:52:55.053028 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:52:55 crc kubenswrapper[4688]: I1001 15:52:55.053926 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.052959 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.053588 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.053649 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.054686 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.054887 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe" gracePeriod=600 Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.929504 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe" exitCode=0 Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.929555 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe"} Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.930020 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced"} Oct 01 15:53:25 crc kubenswrapper[4688]: I1001 15:53:25.930045 4688 scope.go:117] "RemoveContainer" containerID="1c6b7a090cc79a956f03997a2200c6199fe3a354e768ce8f602690ad11f18ba5" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.479234 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6q78r"] Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.482011 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.502221 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6q78r"] Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617642 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpnrk\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-kube-api-access-zpnrk\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617735 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617760 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fc2b544-bb15-4bc0-820e-21cc31e8463a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617781 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fc2b544-bb15-4bc0-820e-21cc31e8463a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617822 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-trusted-ca\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617843 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-bound-sa-token\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617863 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-certificates\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.617878 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-tls\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.639223 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.718735 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fc2b544-bb15-4bc0-820e-21cc31e8463a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.718835 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fc2b544-bb15-4bc0-820e-21cc31e8463a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.718930 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-trusted-ca\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.718972 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-bound-sa-token\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.719018 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-certificates\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.719227 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fc2b544-bb15-4bc0-820e-21cc31e8463a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.721033 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-trusted-ca\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.721231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-certificates\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.721341 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-tls\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.721633 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpnrk\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-kube-api-access-zpnrk\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.725803 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fc2b544-bb15-4bc0-820e-21cc31e8463a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.725997 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-registry-tls\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.738006 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-bound-sa-token\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.743792 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpnrk\" (UniqueName: \"kubernetes.io/projected/3fc2b544-bb15-4bc0-820e-21cc31e8463a-kube-api-access-zpnrk\") pod \"image-registry-66df7c8f76-6q78r\" (UID: \"3fc2b544-bb15-4bc0-820e-21cc31e8463a\") " pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:26 crc kubenswrapper[4688]: I1001 15:54:26.815962 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:27 crc kubenswrapper[4688]: I1001 15:54:27.004600 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6q78r"] Oct 01 15:54:27 crc kubenswrapper[4688]: W1001 15:54:27.012965 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fc2b544_bb15_4bc0_820e_21cc31e8463a.slice/crio-a02881d74894df13aec45be1216a32327b9ad0ede3df04bd77a3123416692b9e WatchSource:0}: Error finding container a02881d74894df13aec45be1216a32327b9ad0ede3df04bd77a3123416692b9e: Status 404 returned error can't find the container with id a02881d74894df13aec45be1216a32327b9ad0ede3df04bd77a3123416692b9e Oct 01 15:54:27 crc kubenswrapper[4688]: I1001 15:54:27.378204 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" event={"ID":"3fc2b544-bb15-4bc0-820e-21cc31e8463a","Type":"ContainerStarted","Data":"c97b1f6603f9890a4541df82ae08a88acdffd8da0df05df0620a7aaeece81609"} Oct 01 15:54:27 crc kubenswrapper[4688]: I1001 15:54:27.378242 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" event={"ID":"3fc2b544-bb15-4bc0-820e-21cc31e8463a","Type":"ContainerStarted","Data":"a02881d74894df13aec45be1216a32327b9ad0ede3df04bd77a3123416692b9e"} Oct 01 15:54:27 crc kubenswrapper[4688]: I1001 15:54:27.379019 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:27 crc kubenswrapper[4688]: I1001 15:54:27.407096 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" podStartSLOduration=1.407076468 podStartE2EDuration="1.407076468s" podCreationTimestamp="2025-10-01 15:54:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:54:27.402958993 +0000 UTC m=+456.753598955" watchObservedRunningTime="2025-10-01 15:54:27.407076468 +0000 UTC m=+456.757716430" Oct 01 15:54:46 crc kubenswrapper[4688]: I1001 15:54:46.824661 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6q78r" Oct 01 15:54:46 crc kubenswrapper[4688]: I1001 15:54:46.885471 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:55:11 crc kubenswrapper[4688]: I1001 15:55:11.924594 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" podUID="761ff908-3503-4509-9fc7-b237a76e3262" containerName="registry" containerID="cri-o://37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08" gracePeriod=30 Oct 01 15:55:11 crc kubenswrapper[4688]: I1001 15:55:11.981064 4688 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-j7jkr container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.20:5000/healthz\": dial tcp 10.217.0.20:5000: connect: connection refused" start-of-body= Oct 01 15:55:11 crc kubenswrapper[4688]: I1001 15:55:11.981122 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" podUID="761ff908-3503-4509-9fc7-b237a76e3262" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.20:5000/healthz\": dial tcp 10.217.0.20:5000: connect: connection refused" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.256567 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.342684 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.343590 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.343866 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.344098 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.344334 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.344597 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.344918 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.345090 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlrxr\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr\") pod \"761ff908-3503-4509-9fc7-b237a76e3262\" (UID: \"761ff908-3503-4509-9fc7-b237a76e3262\") " Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.347717 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.348555 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.350188 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.351042 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.352195 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr" (OuterVolumeSpecName: "kube-api-access-xlrxr") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "kube-api-access-xlrxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.358869 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.358896 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.380685 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "761ff908-3503-4509-9fc7-b237a76e3262" (UID: "761ff908-3503-4509-9fc7-b237a76e3262"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447455 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447761 4688 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447779 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlrxr\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-kube-api-access-xlrxr\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447791 4688 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/761ff908-3503-4509-9fc7-b237a76e3262-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447806 4688 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/761ff908-3503-4509-9fc7-b237a76e3262-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447844 4688 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/761ff908-3503-4509-9fc7-b237a76e3262-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.447855 4688 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/761ff908-3503-4509-9fc7-b237a76e3262-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.691616 4688 generic.go:334] "Generic (PLEG): container finished" podID="761ff908-3503-4509-9fc7-b237a76e3262" containerID="37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08" exitCode=0 Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.691757 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" event={"ID":"761ff908-3503-4509-9fc7-b237a76e3262","Type":"ContainerDied","Data":"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08"} Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.691883 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" event={"ID":"761ff908-3503-4509-9fc7-b237a76e3262","Type":"ContainerDied","Data":"5919a9742fa115faf19277dfb22a82634840a2b950883486239e7730ae90c0d4"} Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.691923 4688 scope.go:117] "RemoveContainer" containerID="37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.691774 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-j7jkr" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.717760 4688 scope.go:117] "RemoveContainer" containerID="37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08" Oct 01 15:55:12 crc kubenswrapper[4688]: E1001 15:55:12.719087 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08\": container with ID starting with 37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08 not found: ID does not exist" containerID="37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.719135 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08"} err="failed to get container status \"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08\": rpc error: code = NotFound desc = could not find container \"37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08\": container with ID starting with 37e57f4ca34904cd23bd38fbaff116a2d99a7208628f7ca8cdf9772a0fea9b08 not found: ID does not exist" Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.738107 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:55:12 crc kubenswrapper[4688]: I1001 15:55:12.744520 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-j7jkr"] Oct 01 15:55:13 crc kubenswrapper[4688]: I1001 15:55:13.389685 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="761ff908-3503-4509-9fc7-b237a76e3262" path="/var/lib/kubelet/pods/761ff908-3503-4509-9fc7-b237a76e3262/volumes" Oct 01 15:55:25 crc kubenswrapper[4688]: I1001 15:55:25.053604 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:55:25 crc kubenswrapper[4688]: I1001 15:55:25.054300 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:55:55 crc kubenswrapper[4688]: I1001 15:55:55.052644 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:55:55 crc kubenswrapper[4688]: I1001 15:55:55.053328 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.052642 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.053238 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.053300 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.054069 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.054166 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced" gracePeriod=600 Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.219376 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced" exitCode=0 Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.219430 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced"} Oct 01 15:56:25 crc kubenswrapper[4688]: I1001 15:56:25.219474 4688 scope.go:117] "RemoveContainer" containerID="23d6c910e6a59dc684eb0dc7ba7bc951105ca05d16bf1358b1bbbf926d8708fe" Oct 01 15:56:26 crc kubenswrapper[4688]: I1001 15:56:26.230399 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94"} Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.064444 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wb49t"] Oct 01 15:56:52 crc kubenswrapper[4688]: E1001 15:56:52.065507 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761ff908-3503-4509-9fc7-b237a76e3262" containerName="registry" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.065539 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="761ff908-3503-4509-9fc7-b237a76e3262" containerName="registry" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.065657 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="761ff908-3503-4509-9fc7-b237a76e3262" containerName="registry" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.066136 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.069350 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.069623 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.078394 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ndwpx"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.079218 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ndwpx" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.080163 4688 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jk5tq" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.090633 4688 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-w7nxz" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.096406 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ndwpx"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.111101 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zxblg"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.111731 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.114543 4688 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qw85l" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.131847 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zxblg"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.133963 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wb49t"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.226280 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgzqk\" (UniqueName: \"kubernetes.io/projected/6654a118-5437-4fb0-9c11-1ac4cf9dba79-kube-api-access-wgzqk\") pod \"cert-manager-webhook-5655c58dd6-zxblg\" (UID: \"6654a118-5437-4fb0-9c11-1ac4cf9dba79\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.226317 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxt7f\" (UniqueName: \"kubernetes.io/projected/6a722b12-5da2-4154-af09-c291e4530375-kube-api-access-mxt7f\") pod \"cert-manager-5b446d88c5-ndwpx\" (UID: \"6a722b12-5da2-4154-af09-c291e4530375\") " pod="cert-manager/cert-manager-5b446d88c5-ndwpx" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.226361 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7bqd\" (UniqueName: \"kubernetes.io/projected/4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f-kube-api-access-w7bqd\") pod \"cert-manager-cainjector-7f985d654d-wb49t\" (UID: \"4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.327450 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7bqd\" (UniqueName: \"kubernetes.io/projected/4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f-kube-api-access-w7bqd\") pod \"cert-manager-cainjector-7f985d654d-wb49t\" (UID: \"4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.327550 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgzqk\" (UniqueName: \"kubernetes.io/projected/6654a118-5437-4fb0-9c11-1ac4cf9dba79-kube-api-access-wgzqk\") pod \"cert-manager-webhook-5655c58dd6-zxblg\" (UID: \"6654a118-5437-4fb0-9c11-1ac4cf9dba79\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.327568 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxt7f\" (UniqueName: \"kubernetes.io/projected/6a722b12-5da2-4154-af09-c291e4530375-kube-api-access-mxt7f\") pod \"cert-manager-5b446d88c5-ndwpx\" (UID: \"6a722b12-5da2-4154-af09-c291e4530375\") " pod="cert-manager/cert-manager-5b446d88c5-ndwpx" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.361267 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7bqd\" (UniqueName: \"kubernetes.io/projected/4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f-kube-api-access-w7bqd\") pod \"cert-manager-cainjector-7f985d654d-wb49t\" (UID: \"4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.375700 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgzqk\" (UniqueName: \"kubernetes.io/projected/6654a118-5437-4fb0-9c11-1ac4cf9dba79-kube-api-access-wgzqk\") pod \"cert-manager-webhook-5655c58dd6-zxblg\" (UID: \"6654a118-5437-4fb0-9c11-1ac4cf9dba79\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.378623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxt7f\" (UniqueName: \"kubernetes.io/projected/6a722b12-5da2-4154-af09-c291e4530375-kube-api-access-mxt7f\") pod \"cert-manager-5b446d88c5-ndwpx\" (UID: \"6a722b12-5da2-4154-af09-c291e4530375\") " pod="cert-manager/cert-manager-5b446d88c5-ndwpx" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.383997 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.391479 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ndwpx" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.422716 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.675565 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ndwpx"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.683687 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.938050 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wb49t"] Oct 01 15:56:52 crc kubenswrapper[4688]: I1001 15:56:52.943254 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-zxblg"] Oct 01 15:56:52 crc kubenswrapper[4688]: W1001 15:56:52.943747 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d4e2651_c2eb_4b0f_a3c1_aec4f0d9c95f.slice/crio-379970c220fc532ef6deb5b019ff12e5a6fdf173261d625adba885d0d0a628a2 WatchSource:0}: Error finding container 379970c220fc532ef6deb5b019ff12e5a6fdf173261d625adba885d0d0a628a2: Status 404 returned error can't find the container with id 379970c220fc532ef6deb5b019ff12e5a6fdf173261d625adba885d0d0a628a2 Oct 01 15:56:53 crc kubenswrapper[4688]: I1001 15:56:53.407687 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ndwpx" event={"ID":"6a722b12-5da2-4154-af09-c291e4530375","Type":"ContainerStarted","Data":"602a5c257912d8ef51cdc91de162c81aa9d3508e7514695fe4481a1bc5e2b0bc"} Oct 01 15:56:53 crc kubenswrapper[4688]: I1001 15:56:53.410211 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" event={"ID":"6654a118-5437-4fb0-9c11-1ac4cf9dba79","Type":"ContainerStarted","Data":"55b21e20f9ac75429d253fd61f66c41942eca8be8ed0cb95d436563f98027c36"} Oct 01 15:56:53 crc kubenswrapper[4688]: I1001 15:56:53.411515 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" event={"ID":"4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f","Type":"ContainerStarted","Data":"379970c220fc532ef6deb5b019ff12e5a6fdf173261d625adba885d0d0a628a2"} Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.435274 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" event={"ID":"6654a118-5437-4fb0-9c11-1ac4cf9dba79","Type":"ContainerStarted","Data":"24df627821d672fd65d05961793799d6bbb86feb53e8e5b3b464f5a719ed95bf"} Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.435813 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.436674 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" event={"ID":"4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f","Type":"ContainerStarted","Data":"e4d559854ba10f9c9f6a0837f5c354509a2c6b925c492d262286ca5f2e30ed97"} Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.438456 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ndwpx" event={"ID":"6a722b12-5da2-4154-af09-c291e4530375","Type":"ContainerStarted","Data":"8481ca9b50a1dc1336545f8c1dc321486d68c5b889d3595c6c4869c39e32c40b"} Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.463388 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" podStartSLOduration=1.5129162219999999 podStartE2EDuration="4.463368385s" podCreationTimestamp="2025-10-01 15:56:52 +0000 UTC" firstStartedPulling="2025-10-01 15:56:52.954619311 +0000 UTC m=+602.305259283" lastFinishedPulling="2025-10-01 15:56:55.905071474 +0000 UTC m=+605.255711446" observedRunningTime="2025-10-01 15:56:56.450482551 +0000 UTC m=+605.801122513" watchObservedRunningTime="2025-10-01 15:56:56.463368385 +0000 UTC m=+605.814008347" Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.481103 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-wb49t" podStartSLOduration=1.452806499 podStartE2EDuration="4.481081998s" podCreationTimestamp="2025-10-01 15:56:52 +0000 UTC" firstStartedPulling="2025-10-01 15:56:52.946932199 +0000 UTC m=+602.297572161" lastFinishedPulling="2025-10-01 15:56:55.975207698 +0000 UTC m=+605.325847660" observedRunningTime="2025-10-01 15:56:56.467275578 +0000 UTC m=+605.817915540" watchObservedRunningTime="2025-10-01 15:56:56.481081998 +0000 UTC m=+605.831721960" Oct 01 15:56:56 crc kubenswrapper[4688]: I1001 15:56:56.481348 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-ndwpx" podStartSLOduration=1.255358873 podStartE2EDuration="4.481341836s" podCreationTimestamp="2025-10-01 15:56:52 +0000 UTC" firstStartedPulling="2025-10-01 15:56:52.683353645 +0000 UTC m=+602.033993607" lastFinishedPulling="2025-10-01 15:56:55.909336608 +0000 UTC m=+605.259976570" observedRunningTime="2025-10-01 15:56:56.477902076 +0000 UTC m=+605.828542038" watchObservedRunningTime="2025-10-01 15:56:56.481341836 +0000 UTC m=+605.831981798" Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.427077 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-zxblg" Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.610939 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xgf9s"] Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.616982 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-controller" containerID="cri-o://5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617020 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="nbdb" containerID="cri-o://797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617192 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="northd" containerID="cri-o://4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617270 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617331 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-node" containerID="cri-o://39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617444 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-acl-logging" containerID="cri-o://5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.617462 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="sbdb" containerID="cri-o://48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.664726 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" containerID="cri-o://6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" gracePeriod=30 Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.971033 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/3.log" Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.973090 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovn-acl-logging/0.log" Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.973623 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovn-controller/0.log" Oct 01 15:57:02 crc kubenswrapper[4688]: I1001 15:57:02.973980 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.033905 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxzb4"] Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034155 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="nbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034177 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="nbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034188 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="northd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034196 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="northd" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034208 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034235 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034248 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034257 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034267 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kubecfg-setup" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034276 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kubecfg-setup" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034287 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034295 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034308 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="sbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034318 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="sbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034334 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-acl-logging" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034345 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-acl-logging" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034360 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034370 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034380 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034389 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034401 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-node" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034410 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-node" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034551 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034564 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-node" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034576 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034586 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="northd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034597 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034609 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034621 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034631 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034642 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="sbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034656 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovn-acl-logging" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034667 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="nbdb" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034780 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034791 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.034803 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034812 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.034938 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" containerName="ovnkube-controller" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.037096 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108070 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108125 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108150 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108167 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108172 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108190 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108259 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108295 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108295 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108317 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108341 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108375 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108407 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108430 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108455 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108484 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108539 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108546 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket" (OuterVolumeSpecName: "log-socket") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108584 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108592 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108565 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108575 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8bsc\" (UniqueName: \"kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108621 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108641 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108674 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108713 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108747 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108778 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108801 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108822 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch\") pod \"7628fc05-465a-425f-b13a-995f8d5fa914\" (UID: \"7628fc05-465a-425f-b13a-995f8d5fa914\") " Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108832 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108861 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108882 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108900 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108922 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash" (OuterVolumeSpecName: "host-slash") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.108940 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log" (OuterVolumeSpecName: "node-log") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109172 4688 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109187 4688 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109198 4688 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109208 4688 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109222 4688 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109233 4688 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109245 4688 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109258 4688 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109285 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109291 4688 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109315 4688 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-log-socket\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109336 4688 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7628fc05-465a-425f-b13a-995f8d5fa914-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109355 4688 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109380 4688 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109433 4688 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109459 4688 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-host-slash\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.109489 4688 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-node-log\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.113486 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc" (OuterVolumeSpecName: "kube-api-access-c8bsc") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "kube-api-access-c8bsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.113684 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.124562 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "7628fc05-465a-425f-b13a-995f8d5fa914" (UID: "7628fc05-465a-425f-b13a-995f8d5fa914"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210591 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-var-lib-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210692 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-etc-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210751 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-log-socket\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-netd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210893 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-netns\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.210964 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-systemd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211018 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211061 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-slash\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211126 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211200 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-script-lib\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211279 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-kubelet\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211338 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvjs\" (UniqueName: \"kubernetes.io/projected/c3e2b83e-390f-4a99-9666-d20003e2d84f-kube-api-access-xzvjs\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211385 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-node-log\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211438 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-systemd-units\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211511 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-ovn\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211602 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-config\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211651 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovn-node-metrics-cert\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211706 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-env-overrides\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211799 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-bin\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211860 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.211986 4688 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.212043 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8bsc\" (UniqueName: \"kubernetes.io/projected/7628fc05-465a-425f-b13a-995f8d5fa914-kube-api-access-c8bsc\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.212072 4688 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7628fc05-465a-425f-b13a-995f8d5fa914-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.212097 4688 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7628fc05-465a-425f-b13a-995f8d5fa914-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.312900 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovn-node-metrics-cert\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.312988 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-env-overrides\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313030 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-bin\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313078 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313122 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-var-lib-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-var-lib-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313200 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-bin\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313224 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313424 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-etc-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313555 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-etc-openvswitch\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313637 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-log-socket\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313715 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-log-socket\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313756 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-netd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313841 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-cni-netd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313878 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-netns\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313972 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-netns\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.313990 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-systemd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314022 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-systemd\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314078 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314126 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314162 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-slash\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314216 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314228 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-script-lib\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314260 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-kubelet\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314265 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314300 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvjs\" (UniqueName: \"kubernetes.io/projected/c3e2b83e-390f-4a99-9666-d20003e2d84f-kube-api-access-xzvjs\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314320 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-slash\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314333 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-node-log\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314416 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-host-kubelet\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314424 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-systemd-units\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314474 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-config\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314519 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-ovn\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314678 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-run-ovn\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314371 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-node-log\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314766 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3e2b83e-390f-4a99-9666-d20003e2d84f-systemd-units\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.314816 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-env-overrides\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.315272 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-script-lib\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.315846 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovnkube-config\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.316875 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3e2b83e-390f-4a99-9666-d20003e2d84f-ovn-node-metrics-cert\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.344395 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvjs\" (UniqueName: \"kubernetes.io/projected/c3e2b83e-390f-4a99-9666-d20003e2d84f-kube-api-access-xzvjs\") pod \"ovnkube-node-qxzb4\" (UID: \"c3e2b83e-390f-4a99-9666-d20003e2d84f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.352108 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:03 crc kubenswrapper[4688]: W1001 15:57:03.384845 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3e2b83e_390f_4a99_9666_d20003e2d84f.slice/crio-8308263d3ff950b4bf83587537df3d913e92e9912a70e9ea6282a9e0da97bb75 WatchSource:0}: Error finding container 8308263d3ff950b4bf83587537df3d913e92e9912a70e9ea6282a9e0da97bb75: Status 404 returned error can't find the container with id 8308263d3ff950b4bf83587537df3d913e92e9912a70e9ea6282a9e0da97bb75 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.482013 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"8308263d3ff950b4bf83587537df3d913e92e9912a70e9ea6282a9e0da97bb75"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.485359 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovnkube-controller/3.log" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.488296 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovn-acl-logging/0.log" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.488907 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xgf9s_7628fc05-465a-425f-b13a-995f8d5fa914/ovn-controller/0.log" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489327 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489367 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489378 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489390 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489401 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489412 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" exitCode=0 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489421 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" exitCode=143 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489431 4688 generic.go:334] "Generic (PLEG): container finished" podID="7628fc05-465a-425f-b13a-995f8d5fa914" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" exitCode=143 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489479 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489511 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489552 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489567 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489580 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489596 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489609 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489621 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489629 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489638 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489645 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489652 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489660 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489668 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489676 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489686 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489697 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489706 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489715 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489722 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489730 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489756 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489764 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489773 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489780 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489788 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489797 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489810 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489819 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489878 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489889 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489896 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489904 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489911 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489918 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489925 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489932 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489944 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" event={"ID":"7628fc05-465a-425f-b13a-995f8d5fa914","Type":"ContainerDied","Data":"cc3060133fd8fb15250beefd6ec6f1675e35047ef3ea93e73136f976134b3735"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489957 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489966 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489974 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489982 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489989 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.489996 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490003 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490010 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490017 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490024 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490041 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.490184 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xgf9s" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.496114 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/1.log" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.496904 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/0.log" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.496965 4688 generic.go:334] "Generic (PLEG): container finished" podID="ac2218c2-479a-4efd-9c1b-ed0d3968b06b" containerID="78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4" exitCode=2 Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.497005 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerDied","Data":"78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.497037 4688 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342"} Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.497817 4688 scope.go:117] "RemoveContainer" containerID="78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.498224 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jcfs4_openshift-multus(ac2218c2-479a-4efd-9c1b-ed0d3968b06b)\"" pod="openshift-multus/multus-jcfs4" podUID="ac2218c2-479a-4efd-9c1b-ed0d3968b06b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.528786 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.530395 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xgf9s"] Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.533719 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xgf9s"] Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.554073 4688 scope.go:117] "RemoveContainer" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.596966 4688 scope.go:117] "RemoveContainer" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.643054 4688 scope.go:117] "RemoveContainer" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.662510 4688 scope.go:117] "RemoveContainer" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.674425 4688 scope.go:117] "RemoveContainer" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.688283 4688 scope.go:117] "RemoveContainer" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.700639 4688 scope.go:117] "RemoveContainer" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.713226 4688 scope.go:117] "RemoveContainer" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.722768 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.723123 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723161 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} err="failed to get container status \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723189 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.723425 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": container with ID starting with 51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93 not found: ID does not exist" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723451 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} err="failed to get container status \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": rpc error: code = NotFound desc = could not find container \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": container with ID starting with 51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723467 4688 scope.go:117] "RemoveContainer" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.723762 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": container with ID starting with 48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904 not found: ID does not exist" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723855 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} err="failed to get container status \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": rpc error: code = NotFound desc = could not find container \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": container with ID starting with 48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.723886 4688 scope.go:117] "RemoveContainer" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.724176 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": container with ID starting with 797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489 not found: ID does not exist" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724206 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} err="failed to get container status \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": rpc error: code = NotFound desc = could not find container \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": container with ID starting with 797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724225 4688 scope.go:117] "RemoveContainer" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.724437 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": container with ID starting with 4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0 not found: ID does not exist" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724459 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} err="failed to get container status \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": rpc error: code = NotFound desc = could not find container \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": container with ID starting with 4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724472 4688 scope.go:117] "RemoveContainer" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.724916 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": container with ID starting with 83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721 not found: ID does not exist" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724943 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} err="failed to get container status \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": rpc error: code = NotFound desc = could not find container \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": container with ID starting with 83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.724962 4688 scope.go:117] "RemoveContainer" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.725120 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": container with ID starting with 39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd not found: ID does not exist" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725144 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} err="failed to get container status \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": rpc error: code = NotFound desc = could not find container \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": container with ID starting with 39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725158 4688 scope.go:117] "RemoveContainer" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.725305 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": container with ID starting with 5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5 not found: ID does not exist" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725328 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} err="failed to get container status \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": rpc error: code = NotFound desc = could not find container \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": container with ID starting with 5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725343 4688 scope.go:117] "RemoveContainer" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.725491 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": container with ID starting with 5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97 not found: ID does not exist" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725533 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} err="failed to get container status \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": rpc error: code = NotFound desc = could not find container \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": container with ID starting with 5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725551 4688 scope.go:117] "RemoveContainer" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: E1001 15:57:03.725730 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": container with ID starting with d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b not found: ID does not exist" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725750 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} err="failed to get container status \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": rpc error: code = NotFound desc = could not find container \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": container with ID starting with d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725760 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725929 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} err="failed to get container status \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.725950 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726121 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} err="failed to get container status \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": rpc error: code = NotFound desc = could not find container \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": container with ID starting with 51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726143 4688 scope.go:117] "RemoveContainer" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726303 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} err="failed to get container status \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": rpc error: code = NotFound desc = could not find container \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": container with ID starting with 48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726319 4688 scope.go:117] "RemoveContainer" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726496 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} err="failed to get container status \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": rpc error: code = NotFound desc = could not find container \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": container with ID starting with 797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726533 4688 scope.go:117] "RemoveContainer" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726707 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} err="failed to get container status \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": rpc error: code = NotFound desc = could not find container \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": container with ID starting with 4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726722 4688 scope.go:117] "RemoveContainer" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726874 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} err="failed to get container status \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": rpc error: code = NotFound desc = could not find container \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": container with ID starting with 83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.726897 4688 scope.go:117] "RemoveContainer" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727251 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} err="failed to get container status \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": rpc error: code = NotFound desc = could not find container \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": container with ID starting with 39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727269 4688 scope.go:117] "RemoveContainer" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727423 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} err="failed to get container status \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": rpc error: code = NotFound desc = could not find container \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": container with ID starting with 5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727450 4688 scope.go:117] "RemoveContainer" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727619 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} err="failed to get container status \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": rpc error: code = NotFound desc = could not find container \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": container with ID starting with 5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727643 4688 scope.go:117] "RemoveContainer" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727929 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} err="failed to get container status \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": rpc error: code = NotFound desc = could not find container \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": container with ID starting with d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.727954 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728143 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} err="failed to get container status \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728162 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728365 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} err="failed to get container status \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": rpc error: code = NotFound desc = could not find container \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": container with ID starting with 51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728402 4688 scope.go:117] "RemoveContainer" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728626 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} err="failed to get container status \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": rpc error: code = NotFound desc = could not find container \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": container with ID starting with 48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728647 4688 scope.go:117] "RemoveContainer" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728865 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} err="failed to get container status \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": rpc error: code = NotFound desc = could not find container \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": container with ID starting with 797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.728890 4688 scope.go:117] "RemoveContainer" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729084 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} err="failed to get container status \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": rpc error: code = NotFound desc = could not find container \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": container with ID starting with 4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729118 4688 scope.go:117] "RemoveContainer" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729284 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} err="failed to get container status \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": rpc error: code = NotFound desc = could not find container \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": container with ID starting with 83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729304 4688 scope.go:117] "RemoveContainer" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729448 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} err="failed to get container status \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": rpc error: code = NotFound desc = could not find container \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": container with ID starting with 39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729465 4688 scope.go:117] "RemoveContainer" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729618 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} err="failed to get container status \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": rpc error: code = NotFound desc = could not find container \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": container with ID starting with 5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729637 4688 scope.go:117] "RemoveContainer" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729817 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} err="failed to get container status \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": rpc error: code = NotFound desc = could not find container \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": container with ID starting with 5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729836 4688 scope.go:117] "RemoveContainer" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729973 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} err="failed to get container status \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": rpc error: code = NotFound desc = could not find container \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": container with ID starting with d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.729991 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730120 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} err="failed to get container status \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730136 4688 scope.go:117] "RemoveContainer" containerID="51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730305 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93"} err="failed to get container status \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": rpc error: code = NotFound desc = could not find container \"51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93\": container with ID starting with 51b5e6a4b1c1af606ca28b362b8183c306fe159e6577c00f0d887a85d7a28c93 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730333 4688 scope.go:117] "RemoveContainer" containerID="48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730599 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904"} err="failed to get container status \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": rpc error: code = NotFound desc = could not find container \"48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904\": container with ID starting with 48dc6ef478615acd89f65e326c974363ee525610b1631276d992412c3d2be904 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730623 4688 scope.go:117] "RemoveContainer" containerID="797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730810 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489"} err="failed to get container status \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": rpc error: code = NotFound desc = could not find container \"797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489\": container with ID starting with 797154e0a0ec9c4246dd8d209feafe24bac170b229c432e404dbc2021d1e0489 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.730827 4688 scope.go:117] "RemoveContainer" containerID="4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731072 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0"} err="failed to get container status \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": rpc error: code = NotFound desc = could not find container \"4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0\": container with ID starting with 4a50f08fbb660a229a6dac873e33567158822655048ab8ff5739498d60a64fb0 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731116 4688 scope.go:117] "RemoveContainer" containerID="83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731304 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721"} err="failed to get container status \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": rpc error: code = NotFound desc = could not find container \"83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721\": container with ID starting with 83fd838a40abe978bde74e884b560d5238f661d695e272fabd52e8879cdc4721 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731323 4688 scope.go:117] "RemoveContainer" containerID="39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731507 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd"} err="failed to get container status \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": rpc error: code = NotFound desc = could not find container \"39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd\": container with ID starting with 39793546aadac42c337a4c88664b53d2a4eaf8f465a53587feaf1b72259da9fd not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731552 4688 scope.go:117] "RemoveContainer" containerID="5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731827 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5"} err="failed to get container status \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": rpc error: code = NotFound desc = could not find container \"5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5\": container with ID starting with 5817ff93799f00efc1afa9185b7a45bd639b525c8e356050bf80feb34f7972d5 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.731853 4688 scope.go:117] "RemoveContainer" containerID="5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.732203 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97"} err="failed to get container status \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": rpc error: code = NotFound desc = could not find container \"5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97\": container with ID starting with 5943e50643b6be224213fbb21f534d52ba7b229d3ba63ac5a495a104f86c9e97 not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.732239 4688 scope.go:117] "RemoveContainer" containerID="d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.732441 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b"} err="failed to get container status \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": rpc error: code = NotFound desc = could not find container \"d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b\": container with ID starting with d25ef89ff7a561f2d3d1b2d0ebfa5f44a50a2327f5ebd15712db413d30bed71b not found: ID does not exist" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.732461 4688 scope.go:117] "RemoveContainer" containerID="6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a" Oct 01 15:57:03 crc kubenswrapper[4688]: I1001 15:57:03.741757 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a"} err="failed to get container status \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": rpc error: code = NotFound desc = could not find container \"6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a\": container with ID starting with 6d89331802f20761eb408a0893a87600d62240f03f3d46e0044e8e47e108e75a not found: ID does not exist" Oct 01 15:57:04 crc kubenswrapper[4688]: I1001 15:57:04.503443 4688 generic.go:334] "Generic (PLEG): container finished" podID="c3e2b83e-390f-4a99-9666-d20003e2d84f" containerID="889f51c094fb23467adaf47e3c96aeddd3048d9cd2c4ca0739314632e04ccf66" exitCode=0 Oct 01 15:57:04 crc kubenswrapper[4688]: I1001 15:57:04.503507 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerDied","Data":"889f51c094fb23467adaf47e3c96aeddd3048d9cd2c4ca0739314632e04ccf66"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.388840 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7628fc05-465a-425f-b13a-995f8d5fa914" path="/var/lib/kubelet/pods/7628fc05-465a-425f-b13a-995f8d5fa914/volumes" Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517027 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"85f349af06f9bea71264d7c38f2fb2b1182f04f34e915804bb391a341c09fe68"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517077 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"f9571f31f90a92f90adac2afcfdcb680737798e2b7e08d2620c76a5a8b30c434"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517090 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"1075e9ab20bc74dd3f3fe97bc169aec2631547fc0fc64e3a3f2dccddddaa3c17"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517101 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"acef2bec5e5be107b736a778af97b9707a97ca933f94b6b04c2386d846470a1b"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517125 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"e569f7586bade3cb7dc41dbfc7ca695cdb5f02e1b4cb99e4712f980e49bac2d9"} Oct 01 15:57:05 crc kubenswrapper[4688]: I1001 15:57:05.517136 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"1579e0edab251b4d7ff62814527a9599848044334866a2e306a30c4c31aaa946"} Oct 01 15:57:08 crc kubenswrapper[4688]: I1001 15:57:08.540613 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"004af31c1a50f87842cfb92391d19400ef0ba7bf42e49be7f8065b7c20b2f373"} Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.555859 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" event={"ID":"c3e2b83e-390f-4a99-9666-d20003e2d84f","Type":"ContainerStarted","Data":"ad97d10a70ca01952c1ac89405a69acb690442da08c5f061faccb8b9e3e55736"} Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.556515 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.556559 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.556572 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.586377 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.587741 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:10 crc kubenswrapper[4688]: I1001 15:57:10.591084 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" podStartSLOduration=7.591073225 podStartE2EDuration="7.591073225s" podCreationTimestamp="2025-10-01 15:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:57:10.589546401 +0000 UTC m=+619.940186373" watchObservedRunningTime="2025-10-01 15:57:10.591073225 +0000 UTC m=+619.941713197" Oct 01 15:57:14 crc kubenswrapper[4688]: I1001 15:57:14.380952 4688 scope.go:117] "RemoveContainer" containerID="78e7d45786fd22901c7efcea31853739a91969211c5ed30730adb47930dbe0e4" Oct 01 15:57:15 crc kubenswrapper[4688]: I1001 15:57:15.587033 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/1.log" Oct 01 15:57:15 crc kubenswrapper[4688]: I1001 15:57:15.588498 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/0.log" Oct 01 15:57:15 crc kubenswrapper[4688]: I1001 15:57:15.588661 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcfs4" event={"ID":"ac2218c2-479a-4efd-9c1b-ed0d3968b06b","Type":"ContainerStarted","Data":"c18ec968c2da0d56fd5cb6d252c54520b0a8fc46b05ce6a020fa03f45e7c948d"} Oct 01 15:57:33 crc kubenswrapper[4688]: I1001 15:57:33.397408 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxzb4" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.777931 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5"] Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.779838 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.783150 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.792598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.792673 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8469l\" (UniqueName: \"kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.792870 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.797503 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5"] Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.894013 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.894135 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.894170 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8469l\" (UniqueName: \"kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.894954 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.895401 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:45 crc kubenswrapper[4688]: I1001 15:57:45.936293 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8469l\" (UniqueName: \"kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:46 crc kubenswrapper[4688]: I1001 15:57:46.100084 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:46 crc kubenswrapper[4688]: I1001 15:57:46.543305 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5"] Oct 01 15:57:46 crc kubenswrapper[4688]: I1001 15:57:46.788997 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerStarted","Data":"ca009b709b1158bd3fea39d76455f770cffd732ea0172fcf5fac24e52cd6ec57"} Oct 01 15:57:46 crc kubenswrapper[4688]: I1001 15:57:46.790683 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerStarted","Data":"5afe68f627faf7693930c3457074a5721b98f053e9b978b0435ee57b90863a78"} Oct 01 15:57:47 crc kubenswrapper[4688]: I1001 15:57:47.799017 4688 generic.go:334] "Generic (PLEG): container finished" podID="cac09066-7376-46af-805b-8d55c8d4045f" containerID="ca009b709b1158bd3fea39d76455f770cffd732ea0172fcf5fac24e52cd6ec57" exitCode=0 Oct 01 15:57:47 crc kubenswrapper[4688]: I1001 15:57:47.799238 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerDied","Data":"ca009b709b1158bd3fea39d76455f770cffd732ea0172fcf5fac24e52cd6ec57"} Oct 01 15:57:50 crc kubenswrapper[4688]: I1001 15:57:50.828925 4688 generic.go:334] "Generic (PLEG): container finished" podID="cac09066-7376-46af-805b-8d55c8d4045f" containerID="29398a76283eb8504ca28938dafa11b0ac18475d1262c7433034c034668611a7" exitCode=0 Oct 01 15:57:50 crc kubenswrapper[4688]: I1001 15:57:50.828997 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerDied","Data":"29398a76283eb8504ca28938dafa11b0ac18475d1262c7433034c034668611a7"} Oct 01 15:57:51 crc kubenswrapper[4688]: I1001 15:57:51.607048 4688 scope.go:117] "RemoveContainer" containerID="329e41cd69386fdff03298ad6afdb247273d28befb874544b94c7dba26ac0342" Oct 01 15:57:51 crc kubenswrapper[4688]: I1001 15:57:51.840272 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcfs4_ac2218c2-479a-4efd-9c1b-ed0d3968b06b/kube-multus/1.log" Oct 01 15:57:51 crc kubenswrapper[4688]: I1001 15:57:51.844182 4688 generic.go:334] "Generic (PLEG): container finished" podID="cac09066-7376-46af-805b-8d55c8d4045f" containerID="9ec96066b9c48846ef3cc49ddd43938e95f8fb0c42f85dd6856a5a838daa7e92" exitCode=0 Oct 01 15:57:51 crc kubenswrapper[4688]: I1001 15:57:51.844261 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerDied","Data":"9ec96066b9c48846ef3cc49ddd43938e95f8fb0c42f85dd6856a5a838daa7e92"} Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.137007 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.295345 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle\") pod \"cac09066-7376-46af-805b-8d55c8d4045f\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.295411 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util\") pod \"cac09066-7376-46af-805b-8d55c8d4045f\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.295477 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8469l\" (UniqueName: \"kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l\") pod \"cac09066-7376-46af-805b-8d55c8d4045f\" (UID: \"cac09066-7376-46af-805b-8d55c8d4045f\") " Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.297208 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle" (OuterVolumeSpecName: "bundle") pod "cac09066-7376-46af-805b-8d55c8d4045f" (UID: "cac09066-7376-46af-805b-8d55c8d4045f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.303736 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l" (OuterVolumeSpecName: "kube-api-access-8469l") pod "cac09066-7376-46af-805b-8d55c8d4045f" (UID: "cac09066-7376-46af-805b-8d55c8d4045f"). InnerVolumeSpecName "kube-api-access-8469l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.322121 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util" (OuterVolumeSpecName: "util") pod "cac09066-7376-46af-805b-8d55c8d4045f" (UID: "cac09066-7376-46af-805b-8d55c8d4045f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.398712 4688 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.398780 4688 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cac09066-7376-46af-805b-8d55c8d4045f-util\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.398809 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8469l\" (UniqueName: \"kubernetes.io/projected/cac09066-7376-46af-805b-8d55c8d4045f-kube-api-access-8469l\") on node \"crc\" DevicePath \"\"" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.861427 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" event={"ID":"cac09066-7376-46af-805b-8d55c8d4045f","Type":"ContainerDied","Data":"5afe68f627faf7693930c3457074a5721b98f053e9b978b0435ee57b90863a78"} Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.861778 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5afe68f627faf7693930c3457074a5721b98f053e9b978b0435ee57b90863a78" Oct 01 15:57:53 crc kubenswrapper[4688]: I1001 15:57:53.861499 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.333593 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 15:57:57 crc kubenswrapper[4688]: E1001 15:57:57.334076 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="extract" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.334091 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="extract" Oct 01 15:57:57 crc kubenswrapper[4688]: E1001 15:57:57.334104 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="pull" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.334109 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="pull" Oct 01 15:57:57 crc kubenswrapper[4688]: E1001 15:57:57.334121 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="util" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.334129 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="util" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.334285 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="cac09066-7376-46af-805b-8d55c8d4045f" containerName="extract" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.334757 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.340354 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.340500 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.340554 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4xp8j" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.345187 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.349291 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bqk4\" (UniqueName: \"kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4\") pod \"nmstate-operator-5d6f6cfd66-hnffz\" (UID: \"fb9b4322-784f-439f-8164-e168bb743200\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.450216 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bqk4\" (UniqueName: \"kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4\") pod \"nmstate-operator-5d6f6cfd66-hnffz\" (UID: \"fb9b4322-784f-439f-8164-e168bb743200\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.469741 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bqk4\" (UniqueName: \"kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4\") pod \"nmstate-operator-5d6f6cfd66-hnffz\" (UID: \"fb9b4322-784f-439f-8164-e168bb743200\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.651360 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 15:57:57 crc kubenswrapper[4688]: I1001 15:57:57.921288 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 15:57:58 crc kubenswrapper[4688]: I1001 15:57:58.892360 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" event={"ID":"fb9b4322-784f-439f-8164-e168bb743200","Type":"ContainerStarted","Data":"01675c9d6bac212f45dea06757ac20a65b1ec10f61acbea2c385e95efde8963c"} Oct 01 15:58:03 crc kubenswrapper[4688]: I1001 15:58:03.940931 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" event={"ID":"fb9b4322-784f-439f-8164-e168bb743200","Type":"ContainerStarted","Data":"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154"} Oct 01 15:58:03 crc kubenswrapper[4688]: I1001 15:58:03.962998 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" podStartSLOduration=1.255322399 podStartE2EDuration="6.962978074s" podCreationTimestamp="2025-10-01 15:57:57 +0000 UTC" firstStartedPulling="2025-10-01 15:57:57.935063592 +0000 UTC m=+667.285703554" lastFinishedPulling="2025-10-01 15:58:03.642719257 +0000 UTC m=+672.993359229" observedRunningTime="2025-10-01 15:58:03.961238268 +0000 UTC m=+673.311878270" watchObservedRunningTime="2025-10-01 15:58:03.962978074 +0000 UTC m=+673.313618036" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.476210 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.478025 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.480737 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-2kz9s" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.490438 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.514789 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-589l7"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.515812 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.517915 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.518639 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.521643 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.548336 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571427 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571478 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571542 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4zwr\" (UniqueName: \"kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571575 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qdqd\" (UniqueName: \"kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd\") pod \"nmstate-metrics-58fcddf996-fj6km\" (UID: \"b03e32c5-d2eb-4077-a598-06ad31d0284b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571602 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.571757 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrzv\" (UniqueName: \"kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.572036 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.647300 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.647935 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.650304 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mvzjf" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.650774 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.652808 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.663736 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673430 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673498 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673536 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673577 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4zwr\" (UniqueName: \"kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673608 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qdqd\" (UniqueName: \"kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd\") pod \"nmstate-metrics-58fcddf996-fj6km\" (UID: \"b03e32c5-d2eb-4077-a598-06ad31d0284b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673633 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673659 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrzv\" (UniqueName: \"kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.673955 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.675036 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.675071 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.698506 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrzv\" (UniqueName: \"kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv\") pod \"nmstate-handler-589l7\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.702229 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.704029 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qdqd\" (UniqueName: \"kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd\") pod \"nmstate-metrics-58fcddf996-fj6km\" (UID: \"b03e32c5-d2eb-4077-a598-06ad31d0284b\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.709042 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4zwr\" (UniqueName: \"kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr\") pod \"nmstate-webhook-6d689559c5-gzngc\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.774655 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.774719 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.774736 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v65kn\" (UniqueName: \"kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.798853 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.852170 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.857806 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.869401 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-689bbd87fc-9wkdq"] Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.870111 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.876372 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.876425 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v65kn\" (UniqueName: \"kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.876452 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: E1001 15:58:06.876605 4688 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 01 15:58:06 crc kubenswrapper[4688]: E1001 15:58:06.876666 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert podName:2a77c482-2518-4f89-bcd1-1b9c4e616d30 nodeName:}" failed. No retries permitted until 2025-10-01 15:58:07.37664448 +0000 UTC m=+676.727284442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-zpjcz" (UID: "2a77c482-2518-4f89-bcd1-1b9c4e616d30") : secret "plugin-serving-cert" not found Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.878188 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.908424 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v65kn\" (UniqueName: \"kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:06 crc kubenswrapper[4688]: W1001 15:58:06.914199 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33578dc7_7ec3_4972_a650_2202a7f7bfa1.slice/crio-58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22 WatchSource:0}: Error finding container 58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22: Status 404 returned error can't find the container with id 58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22 Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.965407 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-589l7" event={"ID":"33578dc7-7ec3-4972-a650-2202a7f7bfa1","Type":"ContainerStarted","Data":"58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22"} Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981042 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-oauth-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981115 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pmn\" (UniqueName: \"kubernetes.io/projected/56e9a377-4db9-4f3e-a573-6f17146eb28e-kube-api-access-28pmn\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981138 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981163 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-service-ca\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981231 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981256 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-trusted-ca-bundle\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.981307 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-oauth-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:06 crc kubenswrapper[4688]: I1001 15:58:06.984637 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-689bbd87fc-9wkdq"] Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.083971 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.084394 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-trusted-ca-bundle\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.084987 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-oauth-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085063 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-oauth-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085088 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pmn\" (UniqueName: \"kubernetes.io/projected/56e9a377-4db9-4f3e-a573-6f17146eb28e-kube-api-access-28pmn\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085129 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085153 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-service-ca\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085720 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-oauth-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.085877 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.087871 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-trusted-ca-bundle\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.091350 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-serving-cert\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.093473 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/56e9a377-4db9-4f3e-a573-6f17146eb28e-console-oauth-config\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.097083 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/56e9a377-4db9-4f3e-a573-6f17146eb28e-service-ca\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.104316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pmn\" (UniqueName: \"kubernetes.io/projected/56e9a377-4db9-4f3e-a573-6f17146eb28e-kube-api-access-28pmn\") pod \"console-689bbd87fc-9wkdq\" (UID: \"56e9a377-4db9-4f3e-a573-6f17146eb28e\") " pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.164475 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.196986 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.203347 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.392612 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.397038 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-zpjcz\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.399142 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-689bbd87fc-9wkdq"] Oct 01 15:58:07 crc kubenswrapper[4688]: W1001 15:58:07.404906 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56e9a377_4db9_4f3e_a573_6f17146eb28e.slice/crio-ddf1b2f56c07aabf0ce1c788ae3a578cc6c303845b91dbc5d32612b7debc2feb WatchSource:0}: Error finding container ddf1b2f56c07aabf0ce1c788ae3a578cc6c303845b91dbc5d32612b7debc2feb: Status 404 returned error can't find the container with id ddf1b2f56c07aabf0ce1c788ae3a578cc6c303845b91dbc5d32612b7debc2feb Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.567256 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.803000 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 15:58:07 crc kubenswrapper[4688]: W1001 15:58:07.814600 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a77c482_2518_4f89_bcd1_1b9c4e616d30.slice/crio-365f8a944ed52630f9dd84d5e2b35716bbe3342df83682be99f30a77498fd07c WatchSource:0}: Error finding container 365f8a944ed52630f9dd84d5e2b35716bbe3342df83682be99f30a77498fd07c: Status 404 returned error can't find the container with id 365f8a944ed52630f9dd84d5e2b35716bbe3342df83682be99f30a77498fd07c Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.972552 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-689bbd87fc-9wkdq" event={"ID":"56e9a377-4db9-4f3e-a573-6f17146eb28e","Type":"ContainerStarted","Data":"25743502c3217aaa82b7dbc3a2fe0d011fcf113af8ab32ef3f73d6fc819ab539"} Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.972602 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-689bbd87fc-9wkdq" event={"ID":"56e9a377-4db9-4f3e-a573-6f17146eb28e","Type":"ContainerStarted","Data":"ddf1b2f56c07aabf0ce1c788ae3a578cc6c303845b91dbc5d32612b7debc2feb"} Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.973366 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerStarted","Data":"4f464b63dc720eb74068ebc7c08f668cdb3d0d6a224a2f4159fcf7684e883805"} Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.974302 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" event={"ID":"2a77c482-2518-4f89-bcd1-1b9c4e616d30","Type":"ContainerStarted","Data":"365f8a944ed52630f9dd84d5e2b35716bbe3342df83682be99f30a77498fd07c"} Oct 01 15:58:07 crc kubenswrapper[4688]: I1001 15:58:07.976453 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" event={"ID":"41322bce-e40d-47d7-9a5d-cd4ae91eb2be","Type":"ContainerStarted","Data":"4df989b61c2dfdd963caf47df95de0eba7d2f3c6ddc553a9028bb2d883560f6f"} Oct 01 15:58:09 crc kubenswrapper[4688]: I1001 15:58:09.014680 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-689bbd87fc-9wkdq" podStartSLOduration=3.01465424 podStartE2EDuration="3.01465424s" podCreationTimestamp="2025-10-01 15:58:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:58:09.003845982 +0000 UTC m=+678.354485944" watchObservedRunningTime="2025-10-01 15:58:09.01465424 +0000 UTC m=+678.365294202" Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.035132 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerStarted","Data":"43e4d93c5aae71ed0e978c4f32f5f20bc04b9868e2e360f407cd36720955612b"} Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.038150 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" event={"ID":"2a77c482-2518-4f89-bcd1-1b9c4e616d30","Type":"ContainerStarted","Data":"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5"} Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.043163 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" event={"ID":"41322bce-e40d-47d7-9a5d-cd4ae91eb2be","Type":"ContainerStarted","Data":"3b90b919bbe18fe24741e5e9fb4f3ee0885a95b894312f2facc174759da8fcd2"} Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.043384 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.046248 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-589l7" event={"ID":"33578dc7-7ec3-4972-a650-2202a7f7bfa1","Type":"ContainerStarted","Data":"0da36f452c607c8830324d13558bd632b21b2fee48afe5db528270dfe2d95f68"} Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.047010 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.061585 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" podStartSLOduration=2.81083391 podStartE2EDuration="9.061497182s" podCreationTimestamp="2025-10-01 15:58:06 +0000 UTC" firstStartedPulling="2025-10-01 15:58:07.817426239 +0000 UTC m=+677.168066211" lastFinishedPulling="2025-10-01 15:58:14.068089491 +0000 UTC m=+683.418729483" observedRunningTime="2025-10-01 15:58:15.055589287 +0000 UTC m=+684.406229279" watchObservedRunningTime="2025-10-01 15:58:15.061497182 +0000 UTC m=+684.412137164" Oct 01 15:58:15 crc kubenswrapper[4688]: I1001 15:58:15.093331 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" podStartSLOduration=2.321562086 podStartE2EDuration="9.093301193s" podCreationTimestamp="2025-10-01 15:58:06 +0000 UTC" firstStartedPulling="2025-10-01 15:58:07.17434992 +0000 UTC m=+676.524989882" lastFinishedPulling="2025-10-01 15:58:13.946088987 +0000 UTC m=+683.296728989" observedRunningTime="2025-10-01 15:58:15.085227792 +0000 UTC m=+684.435867794" watchObservedRunningTime="2025-10-01 15:58:15.093301193 +0000 UTC m=+684.443941205" Oct 01 15:58:17 crc kubenswrapper[4688]: I1001 15:58:17.204376 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:17 crc kubenswrapper[4688]: I1001 15:58:17.204682 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:17 crc kubenswrapper[4688]: I1001 15:58:17.210516 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:17 crc kubenswrapper[4688]: I1001 15:58:17.226775 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-589l7" podStartSLOduration=4.098953938 podStartE2EDuration="11.226758527s" podCreationTimestamp="2025-10-01 15:58:06 +0000 UTC" firstStartedPulling="2025-10-01 15:58:06.917517732 +0000 UTC m=+676.268157694" lastFinishedPulling="2025-10-01 15:58:14.045322321 +0000 UTC m=+683.395962283" observedRunningTime="2025-10-01 15:58:15.12067169 +0000 UTC m=+684.471311652" watchObservedRunningTime="2025-10-01 15:58:17.226758527 +0000 UTC m=+686.577398489" Oct 01 15:58:18 crc kubenswrapper[4688]: I1001 15:58:18.071844 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-689bbd87fc-9wkdq" Oct 01 15:58:18 crc kubenswrapper[4688]: I1001 15:58:18.142768 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:58:19 crc kubenswrapper[4688]: I1001 15:58:19.074713 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerStarted","Data":"2dec712b03f821c287b68ee0e39424c67309584f4d4c137e179b064ec083f1ed"} Oct 01 15:58:19 crc kubenswrapper[4688]: I1001 15:58:19.094957 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" podStartSLOduration=1.5293717409999998 podStartE2EDuration="13.094937323s" podCreationTimestamp="2025-10-01 15:58:06 +0000 UTC" firstStartedPulling="2025-10-01 15:58:07.21558563 +0000 UTC m=+676.566225592" lastFinishedPulling="2025-10-01 15:58:18.781151172 +0000 UTC m=+688.131791174" observedRunningTime="2025-10-01 15:58:19.091838883 +0000 UTC m=+688.442478845" watchObservedRunningTime="2025-10-01 15:58:19.094937323 +0000 UTC m=+688.445577305" Oct 01 15:58:21 crc kubenswrapper[4688]: I1001 15:58:21.883979 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 15:58:25 crc kubenswrapper[4688]: I1001 15:58:25.053141 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:58:25 crc kubenswrapper[4688]: I1001 15:58:25.053642 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:58:26 crc kubenswrapper[4688]: I1001 15:58:26.866390 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.366009 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb"] Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.368455 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.371675 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.397612 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb"] Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.554338 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g97j\" (UniqueName: \"kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.554490 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.554576 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.656307 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.656389 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.656500 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g97j\" (UniqueName: \"kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.657315 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.657451 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.692967 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g97j\" (UniqueName: \"kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:41 crc kubenswrapper[4688]: I1001 15:58:41.702976 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:42 crc kubenswrapper[4688]: I1001 15:58:42.106230 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb"] Oct 01 15:58:42 crc kubenswrapper[4688]: I1001 15:58:42.218051 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" event={"ID":"45408b58-1f2f-4f6b-a526-500f893245f0","Type":"ContainerStarted","Data":"afab390fe16631142e690d87cf9a36a665987a1a707877dcfaf30d313f95b8f3"} Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.196553 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8r5l7" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" containerID="cri-o://15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f" gracePeriod=15 Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.223659 4688 generic.go:334] "Generic (PLEG): container finished" podID="45408b58-1f2f-4f6b-a526-500f893245f0" containerID="d981cf183e25d9a19733eaa00c279ed0aca10a3733dc229e726c34188dfd52d6" exitCode=0 Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.223713 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" event={"ID":"45408b58-1f2f-4f6b-a526-500f893245f0","Type":"ContainerDied","Data":"d981cf183e25d9a19733eaa00c279ed0aca10a3733dc229e726c34188dfd52d6"} Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.595784 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8r5l7_cbca6023-6b5b-43a6-aee8-c5df2797b649/console/0.log" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.596099 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.784936 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdkt2\" (UniqueName: \"kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785366 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785558 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785642 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785715 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785781 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.785858 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config\") pod \"cbca6023-6b5b-43a6-aee8-c5df2797b649\" (UID: \"cbca6023-6b5b-43a6-aee8-c5df2797b649\") " Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.786144 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config" (OuterVolumeSpecName: "console-config") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.786596 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.786714 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca" (OuterVolumeSpecName: "service-ca") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.786822 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.795703 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2" (OuterVolumeSpecName: "kube-api-access-wdkt2") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "kube-api-access-wdkt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.795772 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.796011 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cbca6023-6b5b-43a6-aee8-c5df2797b649" (UID: "cbca6023-6b5b-43a6-aee8-c5df2797b649"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887482 4688 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887557 4688 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887579 4688 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887596 4688 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887614 4688 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887660 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdkt2\" (UniqueName: \"kubernetes.io/projected/cbca6023-6b5b-43a6-aee8-c5df2797b649-kube-api-access-wdkt2\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:43 crc kubenswrapper[4688]: I1001 15:58:43.887679 4688 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbca6023-6b5b-43a6-aee8-c5df2797b649-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.232405 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8r5l7_cbca6023-6b5b-43a6-aee8-c5df2797b649/console/0.log" Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.232943 4688 generic.go:334] "Generic (PLEG): container finished" podID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerID="15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f" exitCode=2 Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.232992 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5l7" event={"ID":"cbca6023-6b5b-43a6-aee8-c5df2797b649","Type":"ContainerDied","Data":"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f"} Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.233037 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8r5l7" Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.233067 4688 scope.go:117] "RemoveContainer" containerID="15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f" Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.233048 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8r5l7" event={"ID":"cbca6023-6b5b-43a6-aee8-c5df2797b649","Type":"ContainerDied","Data":"d184b22acf5d0584d7ac378b2cca7fb221e52128099ea13db9539651d007ffad"} Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.274698 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.279095 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8r5l7"] Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.298480 4688 scope.go:117] "RemoveContainer" containerID="15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f" Oct 01 15:58:44 crc kubenswrapper[4688]: E1001 15:58:44.299009 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f\": container with ID starting with 15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f not found: ID does not exist" containerID="15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f" Oct 01 15:58:44 crc kubenswrapper[4688]: I1001 15:58:44.299059 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f"} err="failed to get container status \"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f\": rpc error: code = NotFound desc = could not find container \"15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f\": container with ID starting with 15601b899a3703c0765fbba3f1e9c89865cd4c16ad0c4a355966f36c5771802f not found: ID does not exist" Oct 01 15:58:45 crc kubenswrapper[4688]: I1001 15:58:45.241609 4688 generic.go:334] "Generic (PLEG): container finished" podID="45408b58-1f2f-4f6b-a526-500f893245f0" containerID="611ee3a4af1827c0e338a589b1cb7027212fcba509a40812d84ca261ad673991" exitCode=0 Oct 01 15:58:45 crc kubenswrapper[4688]: I1001 15:58:45.241708 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" event={"ID":"45408b58-1f2f-4f6b-a526-500f893245f0","Type":"ContainerDied","Data":"611ee3a4af1827c0e338a589b1cb7027212fcba509a40812d84ca261ad673991"} Oct 01 15:58:45 crc kubenswrapper[4688]: I1001 15:58:45.388206 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" path="/var/lib/kubelet/pods/cbca6023-6b5b-43a6-aee8-c5df2797b649/volumes" Oct 01 15:58:46 crc kubenswrapper[4688]: I1001 15:58:46.259249 4688 generic.go:334] "Generic (PLEG): container finished" podID="45408b58-1f2f-4f6b-a526-500f893245f0" containerID="43971e17f157220a487d34bac5d27e70dfbf5bca3006c9e7915ebcbda6a10bf6" exitCode=0 Oct 01 15:58:46 crc kubenswrapper[4688]: I1001 15:58:46.259414 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" event={"ID":"45408b58-1f2f-4f6b-a526-500f893245f0","Type":"ContainerDied","Data":"43971e17f157220a487d34bac5d27e70dfbf5bca3006c9e7915ebcbda6a10bf6"} Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.559141 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.749489 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util\") pod \"45408b58-1f2f-4f6b-a526-500f893245f0\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.749622 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle\") pod \"45408b58-1f2f-4f6b-a526-500f893245f0\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.749718 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g97j\" (UniqueName: \"kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j\") pod \"45408b58-1f2f-4f6b-a526-500f893245f0\" (UID: \"45408b58-1f2f-4f6b-a526-500f893245f0\") " Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.751693 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle" (OuterVolumeSpecName: "bundle") pod "45408b58-1f2f-4f6b-a526-500f893245f0" (UID: "45408b58-1f2f-4f6b-a526-500f893245f0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.758235 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j" (OuterVolumeSpecName: "kube-api-access-6g97j") pod "45408b58-1f2f-4f6b-a526-500f893245f0" (UID: "45408b58-1f2f-4f6b-a526-500f893245f0"). InnerVolumeSpecName "kube-api-access-6g97j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.852051 4688 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.852108 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g97j\" (UniqueName: \"kubernetes.io/projected/45408b58-1f2f-4f6b-a526-500f893245f0-kube-api-access-6g97j\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.925550 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util" (OuterVolumeSpecName: "util") pod "45408b58-1f2f-4f6b-a526-500f893245f0" (UID: "45408b58-1f2f-4f6b-a526-500f893245f0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:58:47 crc kubenswrapper[4688]: I1001 15:58:47.953713 4688 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45408b58-1f2f-4f6b-a526-500f893245f0-util\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:48 crc kubenswrapper[4688]: I1001 15:58:48.278550 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" event={"ID":"45408b58-1f2f-4f6b-a526-500f893245f0","Type":"ContainerDied","Data":"afab390fe16631142e690d87cf9a36a665987a1a707877dcfaf30d313f95b8f3"} Oct 01 15:58:48 crc kubenswrapper[4688]: I1001 15:58:48.278653 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afab390fe16631142e690d87cf9a36a665987a1a707877dcfaf30d313f95b8f3" Oct 01 15:58:48 crc kubenswrapper[4688]: I1001 15:58:48.278700 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb" Oct 01 15:58:54 crc kubenswrapper[4688]: I1001 15:58:54.472595 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:58:54 crc kubenswrapper[4688]: I1001 15:58:54.473367 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerName="controller-manager" containerID="cri-o://fa9d92515bb4ba2e1293a8c0e8b5fee660e8d70635dfd2e08cdc41ff14e74138" gracePeriod=30 Oct 01 15:58:54 crc kubenswrapper[4688]: I1001 15:58:54.548379 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:58:54 crc kubenswrapper[4688]: I1001 15:58:54.548955 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerName="route-controller-manager" containerID="cri-o://ff72140b58e2bd7ec8cdbcb0e57a1a2a086fc6a448b88cd3c68dff1769304a8c" gracePeriod=30 Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.052982 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.053048 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.363218 4688 generic.go:334] "Generic (PLEG): container finished" podID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerID="fa9d92515bb4ba2e1293a8c0e8b5fee660e8d70635dfd2e08cdc41ff14e74138" exitCode=0 Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.363284 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" event={"ID":"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f","Type":"ContainerDied","Data":"fa9d92515bb4ba2e1293a8c0e8b5fee660e8d70635dfd2e08cdc41ff14e74138"} Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.379330 4688 generic.go:334] "Generic (PLEG): container finished" podID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerID="ff72140b58e2bd7ec8cdbcb0e57a1a2a086fc6a448b88cd3c68dff1769304a8c" exitCode=0 Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.379365 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" event={"ID":"fa9c4e7c-802a-4983-9b9c-607673fb8e94","Type":"ContainerDied","Data":"ff72140b58e2bd7ec8cdbcb0e57a1a2a086fc6a448b88cd3c68dff1769304a8c"} Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.470868 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.538480 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.651196 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca\") pod \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652013 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca" (OuterVolumeSpecName: "client-ca") pod "fa9c4e7c-802a-4983-9b9c-607673fb8e94" (UID: "fa9c4e7c-802a-4983-9b9c-607673fb8e94"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.651247 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config\") pod \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652210 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert\") pod \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652246 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles\") pod \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652283 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert\") pod \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652307 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca\") pod \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652327 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq6qq\" (UniqueName: \"kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq\") pod \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652343 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfntc\" (UniqueName: \"kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc\") pod \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\" (UID: \"fa9c4e7c-802a-4983-9b9c-607673fb8e94\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652402 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config\") pod \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\" (UID: \"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f\") " Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.652598 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.653050 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config" (OuterVolumeSpecName: "config") pod "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" (UID: "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.654089 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca" (OuterVolumeSpecName: "client-ca") pod "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" (UID: "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.654455 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" (UID: "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.654996 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config" (OuterVolumeSpecName: "config") pod "fa9c4e7c-802a-4983-9b9c-607673fb8e94" (UID: "fa9c4e7c-802a-4983-9b9c-607673fb8e94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.659450 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" (UID: "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.660352 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq" (OuterVolumeSpecName: "kube-api-access-dq6qq") pod "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" (UID: "8f4d69e6-c470-48df-9d7a-5b1c6c124e4f"). InnerVolumeSpecName "kube-api-access-dq6qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.679285 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fa9c4e7c-802a-4983-9b9c-607673fb8e94" (UID: "fa9c4e7c-802a-4983-9b9c-607673fb8e94"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.695253 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc" (OuterVolumeSpecName: "kube-api-access-pfntc") pod "fa9c4e7c-802a-4983-9b9c-607673fb8e94" (UID: "fa9c4e7c-802a-4983-9b9c-607673fb8e94"). InnerVolumeSpecName "kube-api-access-pfntc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753773 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753812 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9c4e7c-802a-4983-9b9c-607673fb8e94-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753825 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753837 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753853 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9c4e7c-802a-4983-9b9c-607673fb8e94-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753863 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753874 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq6qq\" (UniqueName: \"kubernetes.io/projected/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f-kube-api-access-dq6qq\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:55 crc kubenswrapper[4688]: I1001 15:58:55.753888 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfntc\" (UniqueName: \"kubernetes.io/projected/fa9c4e7c-802a-4983-9b9c-607673fb8e94-kube-api-access-pfntc\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.086852 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc"] Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087123 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087138 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087157 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerName="controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087165 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerName="controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087181 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="util" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087191 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="util" Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087203 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="pull" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087212 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="pull" Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087226 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="extract" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087234 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="extract" Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.087247 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerName="route-controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087255 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerName="route-controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087370 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="45408b58-1f2f-4f6b-a526-500f893245f0" containerName="extract" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087380 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" containerName="controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087390 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" containerName="route-controller-manager" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087404 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbca6023-6b5b-43a6-aee8-c5df2797b649" containerName="console" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.087873 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.090826 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-ff7b8479f-r5hrp"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.091454 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.106180 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ff7b8479f-r5hrp"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.112928 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.251261 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-ff7b8479f-r5hrp"] Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.251730 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-x2lgf proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" podUID="b3e939dc-4569-4d2c-8488-09fdec0fd0a8" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261272 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261319 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vj5j\" (UniqueName: \"kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261350 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2lgf\" (UniqueName: \"kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261369 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261551 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261633 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261657 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.261728 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.282031 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc"] Oct 01 15:58:56 crc kubenswrapper[4688]: E1001 15:58:56.282645 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-6vj5j serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" podUID="897476c6-eb1f-4109-b7da-3aadaaf36f04" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363412 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363474 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363495 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vj5j\" (UniqueName: \"kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363547 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2lgf\" (UniqueName: \"kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363573 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363611 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363636 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363658 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.363678 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.364657 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.365063 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.365108 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.365198 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.365567 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.367973 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.368130 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.387477 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" event={"ID":"8f4d69e6-c470-48df-9d7a-5b1c6c124e4f","Type":"ContainerDied","Data":"d25027240e9b41d09ad25a43fafd39af01ac6cb708890d55fe2502e677e175a9"} Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.387534 4688 scope.go:117] "RemoveContainer" containerID="fa9d92515bb4ba2e1293a8c0e8b5fee660e8d70635dfd2e08cdc41ff14e74138" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.387664 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xrtrx" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.390335 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2lgf\" (UniqueName: \"kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf\") pod \"controller-manager-ff7b8479f-r5hrp\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.391380 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.391442 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" event={"ID":"fa9c4e7c-802a-4983-9b9c-607673fb8e94","Type":"ContainerDied","Data":"fa0630f6ce9db7f0bdd58c04658a5360e032e67ddfd99937b73433d4078663a2"} Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.391569 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.391455 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.398941 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vj5j\" (UniqueName: \"kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j\") pod \"route-controller-manager-5576dcfc69-sd8nc\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.405156 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.412491 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.417697 4688 scope.go:117] "RemoveContainer" containerID="ff72140b58e2bd7ec8cdbcb0e57a1a2a086fc6a448b88cd3c68dff1769304a8c" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.434283 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.435032 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gb5hh"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.449072 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.452535 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xrtrx"] Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.565901 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert\") pod \"897476c6-eb1f-4109-b7da-3aadaaf36f04\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.565941 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert\") pod \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566002 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config\") pod \"897476c6-eb1f-4109-b7da-3aadaaf36f04\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566033 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vj5j\" (UniqueName: \"kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j\") pod \"897476c6-eb1f-4109-b7da-3aadaaf36f04\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566052 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca\") pod \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566097 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config\") pod \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566126 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2lgf\" (UniqueName: \"kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf\") pod \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566161 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca\") pod \"897476c6-eb1f-4109-b7da-3aadaaf36f04\" (UID: \"897476c6-eb1f-4109-b7da-3aadaaf36f04\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566181 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles\") pod \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\" (UID: \"b3e939dc-4569-4d2c-8488-09fdec0fd0a8\") " Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.566833 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b3e939dc-4569-4d2c-8488-09fdec0fd0a8" (UID: "b3e939dc-4569-4d2c-8488-09fdec0fd0a8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.567381 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b3e939dc-4569-4d2c-8488-09fdec0fd0a8" (UID: "b3e939dc-4569-4d2c-8488-09fdec0fd0a8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.567796 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config" (OuterVolumeSpecName: "config") pod "b3e939dc-4569-4d2c-8488-09fdec0fd0a8" (UID: "b3e939dc-4569-4d2c-8488-09fdec0fd0a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.568143 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config" (OuterVolumeSpecName: "config") pod "897476c6-eb1f-4109-b7da-3aadaaf36f04" (UID: "897476c6-eb1f-4109-b7da-3aadaaf36f04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.568392 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca" (OuterVolumeSpecName: "client-ca") pod "897476c6-eb1f-4109-b7da-3aadaaf36f04" (UID: "897476c6-eb1f-4109-b7da-3aadaaf36f04"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.570881 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf" (OuterVolumeSpecName: "kube-api-access-x2lgf") pod "b3e939dc-4569-4d2c-8488-09fdec0fd0a8" (UID: "b3e939dc-4569-4d2c-8488-09fdec0fd0a8"). InnerVolumeSpecName "kube-api-access-x2lgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.570894 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b3e939dc-4569-4d2c-8488-09fdec0fd0a8" (UID: "b3e939dc-4569-4d2c-8488-09fdec0fd0a8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.571677 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "897476c6-eb1f-4109-b7da-3aadaaf36f04" (UID: "897476c6-eb1f-4109-b7da-3aadaaf36f04"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.576372 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j" (OuterVolumeSpecName: "kube-api-access-6vj5j") pod "897476c6-eb1f-4109-b7da-3aadaaf36f04" (UID: "897476c6-eb1f-4109-b7da-3aadaaf36f04"). InnerVolumeSpecName "kube-api-access-6vj5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667816 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/897476c6-eb1f-4109-b7da-3aadaaf36f04-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667863 4688 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667878 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667891 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vj5j\" (UniqueName: \"kubernetes.io/projected/897476c6-eb1f-4109-b7da-3aadaaf36f04-kube-api-access-6vj5j\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667903 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667914 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-config\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667927 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2lgf\" (UniqueName: \"kubernetes.io/projected/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-kube-api-access-x2lgf\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667940 4688 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/897476c6-eb1f-4109-b7da-3aadaaf36f04-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:56 crc kubenswrapper[4688]: I1001 15:58:56.667952 4688 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b3e939dc-4569-4d2c-8488-09fdec0fd0a8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.388669 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f4d69e6-c470-48df-9d7a-5b1c6c124e4f" path="/var/lib/kubelet/pods/8f4d69e6-c470-48df-9d7a-5b1c6c124e4f/volumes" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.389168 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa9c4e7c-802a-4983-9b9c-607673fb8e94" path="/var/lib/kubelet/pods/fa9c4e7c-802a-4983-9b9c-607673fb8e94/volumes" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.399669 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.399684 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff7b8479f-r5hrp" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.436776 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.437735 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.440437 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.440979 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.441270 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.441429 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.441649 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.441773 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.441775 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.446510 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5576dcfc69-sd8nc"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.458756 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.475576 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-ff7b8479f-r5hrp"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.476937 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-ff7b8479f-r5hrp"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.581237 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-client-ca\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.581284 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4wcx\" (UniqueName: \"kubernetes.io/projected/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-kube-api-access-c4wcx\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.581325 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-serving-cert\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.581465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-config\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.682680 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4wcx\" (UniqueName: \"kubernetes.io/projected/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-kube-api-access-c4wcx\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.682741 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-serving-cert\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.682768 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-config\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.682805 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-client-ca\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.683514 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-client-ca\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.685053 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-config\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.702017 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-serving-cert\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.710407 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4wcx\" (UniqueName: \"kubernetes.io/projected/124e2561-7d4d-4052-ab09-8a2c5e3b05e6-kube-api-access-c4wcx\") pod \"route-controller-manager-58b55f4d44-78j4t\" (UID: \"124e2561-7d4d-4052-ab09-8a2c5e3b05e6\") " pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.756826 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.998143 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 15:58:57 crc kubenswrapper[4688]: I1001 15:58:57.998795 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.001558 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.001852 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.002040 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.002227 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.002367 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-hzxvk" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.045092 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.090872 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvj68\" (UniqueName: \"kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.090930 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.090989 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.194159 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvj68\" (UniqueName: \"kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.194209 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.194268 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.218152 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.223299 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.227744 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvj68\" (UniqueName: \"kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68\") pod \"metallb-operator-controller-manager-6b8fb644cb-dm2ql\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.319880 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.350102 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t"] Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.380138 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf"] Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.381299 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.403204 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.403240 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.403377 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hpbl5" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.429018 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" event={"ID":"124e2561-7d4d-4052-ab09-8a2c5e3b05e6","Type":"ContainerStarted","Data":"a76acb02a9e727fbd0c68cde39f849458be14d03349d49a416285dc22c7e7113"} Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.431618 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf"] Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.504488 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.504549 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.504598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvczd\" (UniqueName: \"kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.613740 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.613805 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.613857 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvczd\" (UniqueName: \"kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.625314 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.626207 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.669342 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvczd\" (UniqueName: \"kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd\") pod \"metallb-operator-webhook-server-57dcd5bbd8-lnttf\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.715804 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:58:58 crc kubenswrapper[4688]: I1001 15:58:58.949667 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.138897 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf"] Oct 01 15:58:59 crc kubenswrapper[4688]: W1001 15:58:59.154823 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53e2aa94_6e98_416b_b112_042f4444755d.slice/crio-bcd1423eafcde5fc921af15702f69773fbda7903e053aba1b037827736827765 WatchSource:0}: Error finding container bcd1423eafcde5fc921af15702f69773fbda7903e053aba1b037827736827765: Status 404 returned error can't find the container with id bcd1423eafcde5fc921af15702f69773fbda7903e053aba1b037827736827765 Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.388419 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897476c6-eb1f-4109-b7da-3aadaaf36f04" path="/var/lib/kubelet/pods/897476c6-eb1f-4109-b7da-3aadaaf36f04/volumes" Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.388911 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e939dc-4569-4d2c-8488-09fdec0fd0a8" path="/var/lib/kubelet/pods/b3e939dc-4569-4d2c-8488-09fdec0fd0a8/volumes" Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.434889 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" event={"ID":"124e2561-7d4d-4052-ab09-8a2c5e3b05e6","Type":"ContainerStarted","Data":"188c52534bde58cc5e2408d1c17838cfeaf809a344b5e72a3b2963cd0811a1de"} Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.435117 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.435879 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" event={"ID":"20694bb0-a1ae-4db2-b856-772db76aafd4","Type":"ContainerStarted","Data":"1990c15486ee25be0c1d2ff0e298378ca548b50eb04dc3023f568b784cca86ec"} Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.436680 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" event={"ID":"53e2aa94-6e98-416b-b112-042f4444755d","Type":"ContainerStarted","Data":"bcd1423eafcde5fc921af15702f69773fbda7903e053aba1b037827736827765"} Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.439717 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" Oct 01 15:58:59 crc kubenswrapper[4688]: I1001 15:58:59.459691 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58b55f4d44-78j4t" podStartSLOduration=3.459674283 podStartE2EDuration="3.459674283s" podCreationTimestamp="2025-10-01 15:58:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:58:59.458806678 +0000 UTC m=+728.809446640" watchObservedRunningTime="2025-10-01 15:58:59.459674283 +0000 UTC m=+728.810314245" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.090201 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-84ddb9476-zdx9v"] Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.090837 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.095794 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.095827 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.095878 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.095911 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.096691 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.096826 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.100686 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84ddb9476-zdx9v"] Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.101439 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.235026 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-config\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.235105 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-client-ca\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.235128 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-proxy-ca-bundles\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.235149 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-serving-cert\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.235183 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzr5q\" (UniqueName: \"kubernetes.io/projected/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-kube-api-access-rzr5q\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.336710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-client-ca\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.336753 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-proxy-ca-bundles\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.336779 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-serving-cert\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.336811 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzr5q\" (UniqueName: \"kubernetes.io/projected/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-kube-api-access-rzr5q\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.336841 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-config\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.337769 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-client-ca\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.338118 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-config\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.338147 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-proxy-ca-bundles\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.344171 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-serving-cert\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.369405 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzr5q\" (UniqueName: \"kubernetes.io/projected/4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3-kube-api-access-rzr5q\") pod \"controller-manager-84ddb9476-zdx9v\" (UID: \"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3\") " pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.407992 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:00 crc kubenswrapper[4688]: I1001 15:59:00.787014 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84ddb9476-zdx9v"] Oct 01 15:59:01 crc kubenswrapper[4688]: I1001 15:59:01.470128 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" event={"ID":"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3","Type":"ContainerStarted","Data":"311e7bbe1bb92616002a5f5071a3aa947c19f62bd856484b5cf478ebcc587980"} Oct 01 15:59:01 crc kubenswrapper[4688]: I1001 15:59:01.470400 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" event={"ID":"4efc6eb7-b8e2-4d5f-9ccf-e55ec32d55a3","Type":"ContainerStarted","Data":"cc8d756e868bd48a84108dc9be9874d6e2c80a4ca999a3b1f45b867341ec930e"} Oct 01 15:59:01 crc kubenswrapper[4688]: I1001 15:59:01.470500 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:01 crc kubenswrapper[4688]: I1001 15:59:01.476511 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" Oct 01 15:59:01 crc kubenswrapper[4688]: I1001 15:59:01.491430 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-84ddb9476-zdx9v" podStartSLOduration=5.491409019 podStartE2EDuration="5.491409019s" podCreationTimestamp="2025-10-01 15:58:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:59:01.49076243 +0000 UTC m=+730.841402402" watchObservedRunningTime="2025-10-01 15:59:01.491409019 +0000 UTC m=+730.842048991" Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.516665 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" event={"ID":"20694bb0-a1ae-4db2-b856-772db76aafd4","Type":"ContainerStarted","Data":"2a6846880ab078151e077a6ff47237c1af3948631dfc707cb26bbab7d09601eb"} Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.517250 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.519721 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" event={"ID":"53e2aa94-6e98-416b-b112-042f4444755d","Type":"ContainerStarted","Data":"ece8ec933b7323a08ea17ba7f71f3a3c261c3cd874f7001a1bc8270670ea5f6d"} Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.519874 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.546102 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" podStartSLOduration=2.871356921 podStartE2EDuration="9.546064036s" podCreationTimestamp="2025-10-01 15:58:57 +0000 UTC" firstStartedPulling="2025-10-01 15:58:58.977889149 +0000 UTC m=+728.328529101" lastFinishedPulling="2025-10-01 15:59:05.652596254 +0000 UTC m=+735.003236216" observedRunningTime="2025-10-01 15:59:06.543289986 +0000 UTC m=+735.893929948" watchObservedRunningTime="2025-10-01 15:59:06.546064036 +0000 UTC m=+735.896704018" Oct 01 15:59:06 crc kubenswrapper[4688]: I1001 15:59:06.570704 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" podStartSLOduration=2.056699563 podStartE2EDuration="8.570686714s" podCreationTimestamp="2025-10-01 15:58:58 +0000 UTC" firstStartedPulling="2025-10-01 15:58:59.160454562 +0000 UTC m=+728.511094524" lastFinishedPulling="2025-10-01 15:59:05.674441703 +0000 UTC m=+735.025081675" observedRunningTime="2025-10-01 15:59:06.568681496 +0000 UTC m=+735.919321458" watchObservedRunningTime="2025-10-01 15:59:06.570686714 +0000 UTC m=+735.921326666" Oct 01 15:59:07 crc kubenswrapper[4688]: I1001 15:59:07.008117 4688 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 15:59:18 crc kubenswrapper[4688]: I1001 15:59:18.722962 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.053418 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.053790 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.053835 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.054387 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.054445 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94" gracePeriod=600 Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.626392 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94" exitCode=0 Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.626468 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94"} Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.627013 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd"} Oct 01 15:59:25 crc kubenswrapper[4688]: I1001 15:59:25.627040 4688 scope.go:117] "RemoveContainer" containerID="9b7b498f35db25395f4cc1ac36104207f710fac46dea941c7684f9c130f7dced" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.344632 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.345598 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.362848 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.388664 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s9q7\" (UniqueName: \"kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.388717 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.388739 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.490287 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s9q7\" (UniqueName: \"kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.490355 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.490391 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.490887 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.490984 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.513673 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s9q7\" (UniqueName: \"kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7\") pod \"community-operators-q9gmd\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:26 crc kubenswrapper[4688]: I1001 15:59:26.662803 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:27 crc kubenswrapper[4688]: I1001 15:59:27.177102 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:27 crc kubenswrapper[4688]: W1001 15:59:27.184864 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee3428d0_780a_4c18_a1fd_7d6e5313f6b3.slice/crio-63c5d31a9266ca6cfa722c8a701193ea2a15e0bfbff4ca5bc745793c8ca39cbc WatchSource:0}: Error finding container 63c5d31a9266ca6cfa722c8a701193ea2a15e0bfbff4ca5bc745793c8ca39cbc: Status 404 returned error can't find the container with id 63c5d31a9266ca6cfa722c8a701193ea2a15e0bfbff4ca5bc745793c8ca39cbc Oct 01 15:59:27 crc kubenswrapper[4688]: I1001 15:59:27.640080 4688 generic.go:334] "Generic (PLEG): container finished" podID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerID="37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274" exitCode=0 Oct 01 15:59:27 crc kubenswrapper[4688]: I1001 15:59:27.640161 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerDied","Data":"37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274"} Oct 01 15:59:27 crc kubenswrapper[4688]: I1001 15:59:27.640433 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerStarted","Data":"63c5d31a9266ca6cfa722c8a701193ea2a15e0bfbff4ca5bc745793c8ca39cbc"} Oct 01 15:59:28 crc kubenswrapper[4688]: I1001 15:59:28.656959 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerStarted","Data":"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a"} Oct 01 15:59:29 crc kubenswrapper[4688]: I1001 15:59:29.664852 4688 generic.go:334] "Generic (PLEG): container finished" podID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerID="f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a" exitCode=0 Oct 01 15:59:29 crc kubenswrapper[4688]: I1001 15:59:29.664943 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerDied","Data":"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a"} Oct 01 15:59:30 crc kubenswrapper[4688]: I1001 15:59:30.673425 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerStarted","Data":"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051"} Oct 01 15:59:30 crc kubenswrapper[4688]: I1001 15:59:30.695440 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q9gmd" podStartSLOduration=2.256410457 podStartE2EDuration="4.695413123s" podCreationTimestamp="2025-10-01 15:59:26 +0000 UTC" firstStartedPulling="2025-10-01 15:59:27.641892103 +0000 UTC m=+756.992532095" lastFinishedPulling="2025-10-01 15:59:30.080894799 +0000 UTC m=+759.431534761" observedRunningTime="2025-10-01 15:59:30.69424667 +0000 UTC m=+760.044886632" watchObservedRunningTime="2025-10-01 15:59:30.695413123 +0000 UTC m=+760.046053145" Oct 01 15:59:36 crc kubenswrapper[4688]: I1001 15:59:36.663750 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:36 crc kubenswrapper[4688]: I1001 15:59:36.664364 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:36 crc kubenswrapper[4688]: I1001 15:59:36.706113 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:36 crc kubenswrapper[4688]: I1001 15:59:36.752350 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:36 crc kubenswrapper[4688]: I1001 15:59:36.932941 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:38 crc kubenswrapper[4688]: I1001 15:59:38.323314 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 15:59:38 crc kubenswrapper[4688]: I1001 15:59:38.722663 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q9gmd" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="registry-server" containerID="cri-o://21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051" gracePeriod=2 Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.092209 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-96tnj"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.094609 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.097890 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.098238 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5z4b7" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.099965 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.114360 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.115156 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.116497 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.129563 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159505 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159577 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159675 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159748 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159794 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p276m\" (UniqueName: \"kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159857 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.159895 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gdk6\" (UniqueName: \"kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.160016 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.254806 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zpk8l"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.255720 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.258740 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.259516 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kj76b" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.260865 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.261076 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262256 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262321 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262357 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p276m\" (UniqueName: \"kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262384 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262405 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gdk6\" (UniqueName: \"kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262466 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262553 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.262575 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.262690 4688 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.262903 4688 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.263169 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.263233 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.263298 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert podName:86cf7b4d-c977-4016-b726-57e5330a40fb nodeName:}" failed. No retries permitted until 2025-10-01 15:59:39.762723482 +0000 UTC m=+769.113363524 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert") pod "frr-k8s-webhook-server-5478bdb765-h4qwp" (UID: "86cf7b4d-c977-4016-b726-57e5330a40fb") : secret "frr-k8s-webhook-server-cert" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.263317 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs podName:a9d78e63-3534-48cc-9df2-b986f6897909 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:39.763307609 +0000 UTC m=+769.113947571 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs") pod "frr-k8s-96tnj" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909") : secret "frr-k8s-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.263338 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.263651 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.264374 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.264420 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.285548 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gdk6\" (UniqueName: \"kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.299908 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-85kzw"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.300998 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.301996 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p276m\" (UniqueName: \"kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.305850 4688 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.324063 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-85kzw"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364123 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vrk4\" (UniqueName: \"kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364166 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364201 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364258 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364283 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg65t\" (UniqueName: \"kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364298 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.364313 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467135 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467229 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467268 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg65t\" (UniqueName: \"kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467287 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467287 4688 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467370 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs podName:4990c197-fc4e-484a-8d36-7990a3f3e68a nodeName:}" failed. No retries permitted until 2025-10-01 15:59:39.96734537 +0000 UTC m=+769.317985332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs") pod "controller-5d688f5ffc-85kzw" (UID: "4990c197-fc4e-484a-8d36-7990a3f3e68a") : secret "controller-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467380 4688 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467428 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs podName:1bef642c-98fe-413c-8a6d-a7a308d76036 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:39.967411142 +0000 UTC m=+769.318051104 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs") pod "speaker-zpk8l" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036") : secret "speaker-certs-secret" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467484 4688 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.467513 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist podName:1bef642c-98fe-413c-8a6d-a7a308d76036 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:39.967504894 +0000 UTC m=+769.318144986 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist") pod "speaker-zpk8l" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036") : secret "metallb-memberlist" not found Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467303 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467740 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vrk4\" (UniqueName: \"kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.467803 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.468380 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.477058 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.505910 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg65t\" (UniqueName: \"kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.523988 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vrk4\" (UniqueName: \"kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.600020 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.670209 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s9q7\" (UniqueName: \"kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7\") pod \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.670334 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content\") pod \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.670380 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities\") pod \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\" (UID: \"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3\") " Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.671386 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities" (OuterVolumeSpecName: "utilities") pod "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" (UID: "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.674245 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7" (OuterVolumeSpecName: "kube-api-access-6s9q7") pod "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" (UID: "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3"). InnerVolumeSpecName "kube-api-access-6s9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.719414 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" (UID: "ee3428d0-780a-4c18-a1fd-7d6e5313f6b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.739324 4688 generic.go:334] "Generic (PLEG): container finished" podID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerID="21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051" exitCode=0 Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.739395 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerDied","Data":"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051"} Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.739429 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9gmd" event={"ID":"ee3428d0-780a-4c18-a1fd-7d6e5313f6b3","Type":"ContainerDied","Data":"63c5d31a9266ca6cfa722c8a701193ea2a15e0bfbff4ca5bc745793c8ca39cbc"} Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.739453 4688 scope.go:117] "RemoveContainer" containerID="21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.739460 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9gmd" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.758501 4688 scope.go:117] "RemoveContainer" containerID="f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.771849 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.771918 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.772704 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.772729 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s9q7\" (UniqueName: \"kubernetes.io/projected/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-kube-api-access-6s9q7\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.772742 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.775628 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") pod \"frr-k8s-96tnj\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.776103 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") pod \"frr-k8s-webhook-server-5478bdb765-h4qwp\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.776201 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.781006 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q9gmd"] Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.789274 4688 scope.go:117] "RemoveContainer" containerID="37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.806551 4688 scope.go:117] "RemoveContainer" containerID="21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.807076 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051\": container with ID starting with 21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051 not found: ID does not exist" containerID="21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.807131 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051"} err="failed to get container status \"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051\": rpc error: code = NotFound desc = could not find container \"21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051\": container with ID starting with 21834956bed139840a9e53fcdbbbebe8f036bd165dd3f57fde1852d799c00051 not found: ID does not exist" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.807162 4688 scope.go:117] "RemoveContainer" containerID="f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.807552 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a\": container with ID starting with f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a not found: ID does not exist" containerID="f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.807582 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a"} err="failed to get container status \"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a\": rpc error: code = NotFound desc = could not find container \"f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a\": container with ID starting with f7a91c53ba67e5e3e8394f22d4123c45535c2ddb3bbd30cf4a0281cfed7f5d3a not found: ID does not exist" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.807601 4688 scope.go:117] "RemoveContainer" containerID="37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.807891 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274\": container with ID starting with 37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274 not found: ID does not exist" containerID="37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.807995 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274"} err="failed to get container status \"37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274\": rpc error: code = NotFound desc = could not find container \"37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274\": container with ID starting with 37c6c71005806317d94376f38b452ba7e49ca6a2718c3fe5b3839892cd074274 not found: ID does not exist" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.975140 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.975455 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.975601 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.975897 4688 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 15:59:39 crc kubenswrapper[4688]: E1001 15:59:39.976015 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist podName:1bef642c-98fe-413c-8a6d-a7a308d76036 nodeName:}" failed. No retries permitted until 2025-10-01 15:59:40.975982397 +0000 UTC m=+770.326622399 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist") pod "speaker-zpk8l" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036") : secret "metallb-memberlist" not found Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.980002 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") pod \"controller-5d688f5ffc-85kzw\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:39 crc kubenswrapper[4688]: I1001 15:59:39.983076 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.011886 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.030344 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.240867 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.530394 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 15:59:40 crc kubenswrapper[4688]: W1001 15:59:40.535345 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86cf7b4d_c977_4016_b726_57e5330a40fb.slice/crio-2017a2efcb03a0217985bf05be879520b7b9b7aff8e77d6432c93166e9a5c212 WatchSource:0}: Error finding container 2017a2efcb03a0217985bf05be879520b7b9b7aff8e77d6432c93166e9a5c212: Status 404 returned error can't find the container with id 2017a2efcb03a0217985bf05be879520b7b9b7aff8e77d6432c93166e9a5c212 Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.655153 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-85kzw"] Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.747798 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-85kzw" event={"ID":"4990c197-fc4e-484a-8d36-7990a3f3e68a","Type":"ContainerStarted","Data":"a2f42530fbd8743344233ecf1f11ed09f8901158c8d612bbe154a6cf86d45bc8"} Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.749443 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" event={"ID":"86cf7b4d-c977-4016-b726-57e5330a40fb","Type":"ContainerStarted","Data":"2017a2efcb03a0217985bf05be879520b7b9b7aff8e77d6432c93166e9a5c212"} Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.750330 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"1638e47cf2b1656251a30986adf4d7476de48183582e697bf26fb630765a85e2"} Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.987966 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:40 crc kubenswrapper[4688]: I1001 15:59:40.994625 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"speaker-zpk8l\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " pod="metallb-system/speaker-zpk8l" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.071399 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zpk8l" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.390229 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" path="/var/lib/kubelet/pods/ee3428d0-780a-4c18-a1fd-7d6e5313f6b3/volumes" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.781143 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-85kzw" event={"ID":"4990c197-fc4e-484a-8d36-7990a3f3e68a","Type":"ContainerStarted","Data":"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc"} Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.781218 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-85kzw" event={"ID":"4990c197-fc4e-484a-8d36-7990a3f3e68a","Type":"ContainerStarted","Data":"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049"} Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.781410 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.795452 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zpk8l" event={"ID":"1bef642c-98fe-413c-8a6d-a7a308d76036","Type":"ContainerStarted","Data":"498d691e8ffbfc6c70c3503591c296a306aecc3c9e063777f18102b8feadf6b4"} Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.795505 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zpk8l" event={"ID":"1bef642c-98fe-413c-8a6d-a7a308d76036","Type":"ContainerStarted","Data":"3fa460f893bab5e197bfe5d4027cdcd0e3713a87924aeb1129495be6e1247a49"} Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.795535 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zpk8l" event={"ID":"1bef642c-98fe-413c-8a6d-a7a308d76036","Type":"ContainerStarted","Data":"e4d230da387cd7935bfce95a4f27e9be8de670d51968278f805722d21c603cfe"} Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.796201 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zpk8l" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.802337 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-85kzw" podStartSLOduration=2.802290612 podStartE2EDuration="2.802290612s" podCreationTimestamp="2025-10-01 15:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:59:41.798462702 +0000 UTC m=+771.149102674" watchObservedRunningTime="2025-10-01 15:59:41.802290612 +0000 UTC m=+771.152930574" Oct 01 15:59:41 crc kubenswrapper[4688]: I1001 15:59:41.839873 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zpk8l" podStartSLOduration=2.839858633 podStartE2EDuration="2.839858633s" podCreationTimestamp="2025-10-01 15:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 15:59:41.836414024 +0000 UTC m=+771.187053986" watchObservedRunningTime="2025-10-01 15:59:41.839858633 +0000 UTC m=+771.190498595" Oct 01 15:59:48 crc kubenswrapper[4688]: I1001 15:59:48.864608 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="5abf2f107eda6674a107691f4534b62321e321947f60b5f7e1b84bfc9dda4719" exitCode=0 Oct 01 15:59:48 crc kubenswrapper[4688]: I1001 15:59:48.864662 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerDied","Data":"5abf2f107eda6674a107691f4534b62321e321947f60b5f7e1b84bfc9dda4719"} Oct 01 15:59:48 crc kubenswrapper[4688]: I1001 15:59:48.867000 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" event={"ID":"86cf7b4d-c977-4016-b726-57e5330a40fb","Type":"ContainerStarted","Data":"5d1ea6733116b3f7ae95799de9b2055125378f9095ff12acaf64e7a10497f1be"} Oct 01 15:59:48 crc kubenswrapper[4688]: I1001 15:59:48.867159 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 15:59:48 crc kubenswrapper[4688]: I1001 15:59:48.909125 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" podStartSLOduration=2.611253285 podStartE2EDuration="9.909107663s" podCreationTimestamp="2025-10-01 15:59:39 +0000 UTC" firstStartedPulling="2025-10-01 15:59:40.537799824 +0000 UTC m=+769.888439786" lastFinishedPulling="2025-10-01 15:59:47.835654202 +0000 UTC m=+777.186294164" observedRunningTime="2025-10-01 15:59:48.904949633 +0000 UTC m=+778.255589615" watchObservedRunningTime="2025-10-01 15:59:48.909107663 +0000 UTC m=+778.259747625" Oct 01 15:59:49 crc kubenswrapper[4688]: I1001 15:59:49.874801 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="ad383a58fb88d5d9811b0ea386b54a8aeb0c6d75709080d2b393b1139ec7eea1" exitCode=0 Oct 01 15:59:49 crc kubenswrapper[4688]: I1001 15:59:49.874892 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerDied","Data":"ad383a58fb88d5d9811b0ea386b54a8aeb0c6d75709080d2b393b1139ec7eea1"} Oct 01 15:59:50 crc kubenswrapper[4688]: I1001 15:59:50.246494 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 15:59:50 crc kubenswrapper[4688]: I1001 15:59:50.884064 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="8e380136f2b94fb2686295d273c4d06283e596f9754b8601e2359f68d39790ab" exitCode=0 Oct 01 15:59:50 crc kubenswrapper[4688]: I1001 15:59:50.884109 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerDied","Data":"8e380136f2b94fb2686295d273c4d06283e596f9754b8601e2359f68d39790ab"} Oct 01 15:59:51 crc kubenswrapper[4688]: I1001 15:59:51.077730 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zpk8l" Oct 01 15:59:51 crc kubenswrapper[4688]: I1001 15:59:51.892930 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"d1c486b572834d5e19902b9da70241b6656ecc730701014b1a3db045765415e5"} Oct 01 15:59:51 crc kubenswrapper[4688]: I1001 15:59:51.893241 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"2adff539b744d26bf47f3bf8345c86fa4837bdf3d800287b224f0985f224248b"} Oct 01 15:59:51 crc kubenswrapper[4688]: I1001 15:59:51.893258 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"cd481bc67fef91619ada7075f2e5a399338400fa039814f49a1679bb1d4845e4"} Oct 01 15:59:51 crc kubenswrapper[4688]: I1001 15:59:51.893269 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"3ce17c0c71b6b890863cac3393333b9816e458d480af5e92d43ff177e0f40505"} Oct 01 15:59:52 crc kubenswrapper[4688]: I1001 15:59:52.905569 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"ca7b92be6bbd3fefceb16c315dbd5a319f5a5109bc38beb42a83d64eb24ef1a4"} Oct 01 15:59:52 crc kubenswrapper[4688]: I1001 15:59:52.906034 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:52 crc kubenswrapper[4688]: I1001 15:59:52.906062 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96tnj" event={"ID":"a9d78e63-3534-48cc-9df2-b986f6897909","Type":"ContainerStarted","Data":"b4e6fc1a1be0dd3cdd6e66cfd1f47397c89fa6d8f46db6a904454ee9f741fa3e"} Oct 01 15:59:52 crc kubenswrapper[4688]: I1001 15:59:52.961297 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-96tnj" podStartSLOduration=6.364973689 podStartE2EDuration="13.961278245s" podCreationTimestamp="2025-10-01 15:59:39 +0000 UTC" firstStartedPulling="2025-10-01 15:59:40.233889359 +0000 UTC m=+769.584529321" lastFinishedPulling="2025-10-01 15:59:47.830193915 +0000 UTC m=+777.180833877" observedRunningTime="2025-10-01 15:59:52.956659141 +0000 UTC m=+782.307299123" watchObservedRunningTime="2025-10-01 15:59:52.961278245 +0000 UTC m=+782.311918207" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.556328 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 15:59:54 crc kubenswrapper[4688]: E1001 15:59:54.556899 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="extract-utilities" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.556915 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="extract-utilities" Oct 01 15:59:54 crc kubenswrapper[4688]: E1001 15:59:54.556929 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="registry-server" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.556936 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="registry-server" Oct 01 15:59:54 crc kubenswrapper[4688]: E1001 15:59:54.556952 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="extract-content" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.556959 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="extract-content" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.557090 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3428d0-780a-4c18-a1fd-7d6e5313f6b3" containerName="registry-server" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.557575 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.562441 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bg89x" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.562689 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.562755 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.575759 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.679972 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxjsx\" (UniqueName: \"kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx\") pod \"openstack-operator-index-2b2z8\" (UID: \"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7\") " pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.781278 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxjsx\" (UniqueName: \"kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx\") pod \"openstack-operator-index-2b2z8\" (UID: \"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7\") " pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.802329 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxjsx\" (UniqueName: \"kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx\") pod \"openstack-operator-index-2b2z8\" (UID: \"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7\") " pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 15:59:54 crc kubenswrapper[4688]: I1001 15:59:54.871803 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 15:59:55 crc kubenswrapper[4688]: I1001 15:59:55.013750 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:55 crc kubenswrapper[4688]: I1001 15:59:55.059073 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-96tnj" Oct 01 15:59:55 crc kubenswrapper[4688]: W1001 15:59:55.123083 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0f76cfe_1f6b_4d0a_9454_b4d722e93dc7.slice/crio-9a6a3b3bdc556ae31f4e7719263fe5cc899a51040268842c4ac5a4b7d394b77c WatchSource:0}: Error finding container 9a6a3b3bdc556ae31f4e7719263fe5cc899a51040268842c4ac5a4b7d394b77c: Status 404 returned error can't find the container with id 9a6a3b3bdc556ae31f4e7719263fe5cc899a51040268842c4ac5a4b7d394b77c Oct 01 15:59:55 crc kubenswrapper[4688]: I1001 15:59:55.129199 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 15:59:55 crc kubenswrapper[4688]: I1001 15:59:55.940590 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2b2z8" event={"ID":"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7","Type":"ContainerStarted","Data":"9a6a3b3bdc556ae31f4e7719263fe5cc899a51040268842c4ac5a4b7d394b77c"} Oct 01 15:59:58 crc kubenswrapper[4688]: I1001 15:59:58.966122 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2b2z8" event={"ID":"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7","Type":"ContainerStarted","Data":"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43"} Oct 01 15:59:58 crc kubenswrapper[4688]: I1001 15:59:58.986141 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2b2z8" podStartSLOduration=1.644792551 podStartE2EDuration="4.986125308s" podCreationTimestamp="2025-10-01 15:59:54 +0000 UTC" firstStartedPulling="2025-10-01 15:59:55.124916274 +0000 UTC m=+784.475556236" lastFinishedPulling="2025-10-01 15:59:58.466249021 +0000 UTC m=+787.816888993" observedRunningTime="2025-10-01 15:59:58.984073568 +0000 UTC m=+788.334713560" watchObservedRunningTime="2025-10-01 15:59:58.986125308 +0000 UTC m=+788.336765270" Oct 01 15:59:59 crc kubenswrapper[4688]: I1001 15:59:59.113854 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 15:59:59 crc kubenswrapper[4688]: I1001 15:59:59.916351 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-9cs6j"] Oct 01 15:59:59 crc kubenswrapper[4688]: I1001 15:59:59.918073 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 15:59:59 crc kubenswrapper[4688]: I1001 15:59:59.925472 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9cs6j"] Oct 01 15:59:59 crc kubenswrapper[4688]: I1001 15:59:59.955827 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74nsl\" (UniqueName: \"kubernetes.io/projected/20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010-kube-api-access-74nsl\") pod \"openstack-operator-index-9cs6j\" (UID: \"20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010\") " pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.040022 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.057094 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74nsl\" (UniqueName: \"kubernetes.io/projected/20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010-kube-api-access-74nsl\") pod \"openstack-operator-index-9cs6j\" (UID: \"20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010\") " pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.119264 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74nsl\" (UniqueName: \"kubernetes.io/projected/20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010-kube-api-access-74nsl\") pod \"openstack-operator-index-9cs6j\" (UID: \"20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010\") " pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.150252 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc"] Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.151108 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.153781 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.154308 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.166463 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc"] Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.261088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.261175 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v767g\" (UniqueName: \"kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.261205 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.262459 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.362674 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v767g\" (UniqueName: \"kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.362730 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.362764 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.363995 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.379311 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.385172 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v767g\" (UniqueName: \"kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g\") pod \"collect-profiles-29322240-ssdlc\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.472136 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.757575 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9cs6j"] Oct 01 16:00:00 crc kubenswrapper[4688]: W1001 16:00:00.767312 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20aa9d1e_9bdb_45fb_8b83_ce5aac3f0010.slice/crio-f9b25d11d5e803e9fdfa7148bde050d3070d83255cb163e71935f634df131646 WatchSource:0}: Error finding container f9b25d11d5e803e9fdfa7148bde050d3070d83255cb163e71935f634df131646: Status 404 returned error can't find the container with id f9b25d11d5e803e9fdfa7148bde050d3070d83255cb163e71935f634df131646 Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.900252 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc"] Oct 01 16:00:00 crc kubenswrapper[4688]: W1001 16:00:00.906601 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39498707_adfc_4c83_a1e3_da36568ff834.slice/crio-47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc WatchSource:0}: Error finding container 47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc: Status 404 returned error can't find the container with id 47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.982954 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9cs6j" event={"ID":"20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010","Type":"ContainerStarted","Data":"f9b25d11d5e803e9fdfa7148bde050d3070d83255cb163e71935f634df131646"} Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.984250 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" event={"ID":"39498707-adfc-4c83-a1e3-da36568ff834","Type":"ContainerStarted","Data":"47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc"} Oct 01 16:00:00 crc kubenswrapper[4688]: I1001 16:00:00.984404 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2b2z8" podUID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" containerName="registry-server" containerID="cri-o://bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43" gracePeriod=2 Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.523424 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.578369 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxjsx\" (UniqueName: \"kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx\") pod \"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7\" (UID: \"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7\") " Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.584410 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx" (OuterVolumeSpecName: "kube-api-access-kxjsx") pod "a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" (UID: "a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7"). InnerVolumeSpecName "kube-api-access-kxjsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.679497 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxjsx\" (UniqueName: \"kubernetes.io/projected/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7-kube-api-access-kxjsx\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.990970 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9cs6j" event={"ID":"20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010","Type":"ContainerStarted","Data":"0aaaa0dbd51a4f8bbb5252dce03a849189ee66dc496afa0e10813e1e02e01718"} Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.992746 4688 generic.go:334] "Generic (PLEG): container finished" podID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" containerID="bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43" exitCode=0 Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.992819 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2b2z8" event={"ID":"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7","Type":"ContainerDied","Data":"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43"} Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.992904 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2b2z8" event={"ID":"a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7","Type":"ContainerDied","Data":"9a6a3b3bdc556ae31f4e7719263fe5cc899a51040268842c4ac5a4b7d394b77c"} Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.992825 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2b2z8" Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.992927 4688 scope.go:117] "RemoveContainer" containerID="bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43" Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.994574 4688 generic.go:334] "Generic (PLEG): container finished" podID="39498707-adfc-4c83-a1e3-da36568ff834" containerID="c6a5bbd0ec084639bcb0a1076a97c7d347ea8ebccfd66c85d592f77fba732b61" exitCode=0 Oct 01 16:00:01 crc kubenswrapper[4688]: I1001 16:00:01.994601 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" event={"ID":"39498707-adfc-4c83-a1e3-da36568ff834","Type":"ContainerDied","Data":"c6a5bbd0ec084639bcb0a1076a97c7d347ea8ebccfd66c85d592f77fba732b61"} Oct 01 16:00:02 crc kubenswrapper[4688]: I1001 16:00:02.008670 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-9cs6j" podStartSLOduration=2.275741118 podStartE2EDuration="3.008647083s" podCreationTimestamp="2025-10-01 15:59:59 +0000 UTC" firstStartedPulling="2025-10-01 16:00:00.771190787 +0000 UTC m=+790.121830749" lastFinishedPulling="2025-10-01 16:00:01.504096742 +0000 UTC m=+790.854736714" observedRunningTime="2025-10-01 16:00:02.008251002 +0000 UTC m=+791.358891004" watchObservedRunningTime="2025-10-01 16:00:02.008647083 +0000 UTC m=+791.359287065" Oct 01 16:00:02 crc kubenswrapper[4688]: I1001 16:00:02.013904 4688 scope.go:117] "RemoveContainer" containerID="bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43" Oct 01 16:00:02 crc kubenswrapper[4688]: E1001 16:00:02.014392 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43\": container with ID starting with bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43 not found: ID does not exist" containerID="bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43" Oct 01 16:00:02 crc kubenswrapper[4688]: I1001 16:00:02.014433 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43"} err="failed to get container status \"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43\": rpc error: code = NotFound desc = could not find container \"bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43\": container with ID starting with bb032529c7b39626392fa9c723bd689db0472cdb435756cfcc3893f2b1483b43 not found: ID does not exist" Oct 01 16:00:02 crc kubenswrapper[4688]: I1001 16:00:02.048320 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 16:00:02 crc kubenswrapper[4688]: I1001 16:00:02.052131 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2b2z8"] Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.304738 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.390702 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" path="/var/lib/kubelet/pods/a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7/volumes" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.412143 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume\") pod \"39498707-adfc-4c83-a1e3-da36568ff834\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.412228 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume\") pod \"39498707-adfc-4c83-a1e3-da36568ff834\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.412269 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v767g\" (UniqueName: \"kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g\") pod \"39498707-adfc-4c83-a1e3-da36568ff834\" (UID: \"39498707-adfc-4c83-a1e3-da36568ff834\") " Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.413159 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume" (OuterVolumeSpecName: "config-volume") pod "39498707-adfc-4c83-a1e3-da36568ff834" (UID: "39498707-adfc-4c83-a1e3-da36568ff834"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.418135 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g" (OuterVolumeSpecName: "kube-api-access-v767g") pod "39498707-adfc-4c83-a1e3-da36568ff834" (UID: "39498707-adfc-4c83-a1e3-da36568ff834"). InnerVolumeSpecName "kube-api-access-v767g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.419728 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "39498707-adfc-4c83-a1e3-da36568ff834" (UID: "39498707-adfc-4c83-a1e3-da36568ff834"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.513884 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/39498707-adfc-4c83-a1e3-da36568ff834-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.513917 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v767g\" (UniqueName: \"kubernetes.io/projected/39498707-adfc-4c83-a1e3-da36568ff834-kube-api-access-v767g\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:03 crc kubenswrapper[4688]: I1001 16:00:03.513929 4688 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/39498707-adfc-4c83-a1e3-da36568ff834-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:04 crc kubenswrapper[4688]: I1001 16:00:04.011569 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" event={"ID":"39498707-adfc-4c83-a1e3-da36568ff834","Type":"ContainerDied","Data":"47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc"} Oct 01 16:00:04 crc kubenswrapper[4688]: I1001 16:00:04.011885 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47f832df2cc9011a14f7093c10f78a4f24efdd0a2b0788e7ff240ff96452b8dc" Oct 01 16:00:04 crc kubenswrapper[4688]: I1001 16:00:04.011698 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.015731 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-96tnj" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.121898 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:10 crc kubenswrapper[4688]: E1001 16:00:10.122262 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" containerName="registry-server" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.122290 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" containerName="registry-server" Oct 01 16:00:10 crc kubenswrapper[4688]: E1001 16:00:10.122317 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39498707-adfc-4c83-a1e3-da36568ff834" containerName="collect-profiles" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.122331 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="39498707-adfc-4c83-a1e3-da36568ff834" containerName="collect-profiles" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.122593 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="39498707-adfc-4c83-a1e3-da36568ff834" containerName="collect-profiles" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.122623 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f76cfe-1f6b-4d0a-9454-b4d722e93dc7" containerName="registry-server" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.124125 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.142797 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.200840 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d424m\" (UniqueName: \"kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.200928 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.201109 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.263335 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.263645 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.293932 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.302018 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.302157 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d424m\" (UniqueName: \"kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.302202 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.302666 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.302871 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.330315 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d424m\" (UniqueName: \"kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m\") pod \"redhat-operators-46vfm\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.445494 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:10 crc kubenswrapper[4688]: I1001 16:00:10.997991 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:11 crc kubenswrapper[4688]: I1001 16:00:11.062229 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerStarted","Data":"4fea486182b2cc94f2c769389539195b87bdf876a3bfdda0b7dcef84f17588e5"} Oct 01 16:00:11 crc kubenswrapper[4688]: I1001 16:00:11.102334 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-9cs6j" Oct 01 16:00:12 crc kubenswrapper[4688]: I1001 16:00:12.073914 4688 generic.go:334] "Generic (PLEG): container finished" podID="a4128300-2382-4b97-a080-249362e80d6e" containerID="6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e" exitCode=0 Oct 01 16:00:12 crc kubenswrapper[4688]: I1001 16:00:12.073990 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerDied","Data":"6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e"} Oct 01 16:00:14 crc kubenswrapper[4688]: I1001 16:00:14.099358 4688 generic.go:334] "Generic (PLEG): container finished" podID="a4128300-2382-4b97-a080-249362e80d6e" containerID="ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c" exitCode=0 Oct 01 16:00:14 crc kubenswrapper[4688]: I1001 16:00:14.099426 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerDied","Data":"ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c"} Oct 01 16:00:14 crc kubenswrapper[4688]: I1001 16:00:14.933384 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:14 crc kubenswrapper[4688]: I1001 16:00:14.946602 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:14 crc kubenswrapper[4688]: I1001 16:00:14.946902 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.070807 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.070882 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq9xs\" (UniqueName: \"kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.070975 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.106566 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerStarted","Data":"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7"} Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.133702 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-46vfm" podStartSLOduration=2.3873852380000002 podStartE2EDuration="5.133683258s" podCreationTimestamp="2025-10-01 16:00:10 +0000 UTC" firstStartedPulling="2025-10-01 16:00:12.077751391 +0000 UTC m=+801.428391393" lastFinishedPulling="2025-10-01 16:00:14.824049411 +0000 UTC m=+804.174689413" observedRunningTime="2025-10-01 16:00:15.131925677 +0000 UTC m=+804.482565659" watchObservedRunningTime="2025-10-01 16:00:15.133683258 +0000 UTC m=+804.484323230" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.172276 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.172622 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.172821 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq9xs\" (UniqueName: \"kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.173600 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.173643 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.203654 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq9xs\" (UniqueName: \"kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs\") pod \"redhat-marketplace-z45sc\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.264366 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:15 crc kubenswrapper[4688]: I1001 16:00:15.696215 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.112810 4688 generic.go:334] "Generic (PLEG): container finished" podID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerID="7474f86f53785dab4b5ec9805e0aebc8a634792ff318211cee2c64bc1e13bb19" exitCode=0 Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.112987 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerDied","Data":"7474f86f53785dab4b5ec9805e0aebc8a634792ff318211cee2c64bc1e13bb19"} Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.113360 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerStarted","Data":"249c84b869597692a53e68ad62386d1cd37329ce6ca9c03093862ef0d8f400c3"} Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.558064 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg"] Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.559412 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.561277 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-tknj4" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.574689 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg"] Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.692036 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.692118 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.692148 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tps94\" (UniqueName: \"kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.793746 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.793814 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tps94\" (UniqueName: \"kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.793921 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.794769 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.794783 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.824727 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tps94\" (UniqueName: \"kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94\") pod \"732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:16 crc kubenswrapper[4688]: I1001 16:00:16.890172 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:17 crc kubenswrapper[4688]: I1001 16:00:17.119320 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerStarted","Data":"49f49eb4fb0cff420e54ff2ee89dd5a7bf8b1456b10e2ee6179ca580ba7b3e6e"} Oct 01 16:00:17 crc kubenswrapper[4688]: I1001 16:00:17.396702 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg"] Oct 01 16:00:18 crc kubenswrapper[4688]: I1001 16:00:18.126202 4688 generic.go:334] "Generic (PLEG): container finished" podID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerID="49f49eb4fb0cff420e54ff2ee89dd5a7bf8b1456b10e2ee6179ca580ba7b3e6e" exitCode=0 Oct 01 16:00:18 crc kubenswrapper[4688]: I1001 16:00:18.126561 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerDied","Data":"49f49eb4fb0cff420e54ff2ee89dd5a7bf8b1456b10e2ee6179ca580ba7b3e6e"} Oct 01 16:00:18 crc kubenswrapper[4688]: I1001 16:00:18.129445 4688 generic.go:334] "Generic (PLEG): container finished" podID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerID="9a1036bf476abc7aa6487d88b60feb199d93c9bd445587d669392a25ba6ba366" exitCode=0 Oct 01 16:00:18 crc kubenswrapper[4688]: I1001 16:00:18.129473 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerDied","Data":"9a1036bf476abc7aa6487d88b60feb199d93c9bd445587d669392a25ba6ba366"} Oct 01 16:00:18 crc kubenswrapper[4688]: I1001 16:00:18.129490 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerStarted","Data":"732ca1d85fc300087df2583aee1c59abeb4dd73fb7cca08fd4f435b229b3425c"} Oct 01 16:00:19 crc kubenswrapper[4688]: I1001 16:00:19.140305 4688 generic.go:334] "Generic (PLEG): container finished" podID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerID="08ed5a426f720a605495a3eade467039182ae959f5ebf7b3b657bd091ec7469c" exitCode=0 Oct 01 16:00:19 crc kubenswrapper[4688]: I1001 16:00:19.140432 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerDied","Data":"08ed5a426f720a605495a3eade467039182ae959f5ebf7b3b657bd091ec7469c"} Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.151770 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerStarted","Data":"b63d98fc79af68824e8387a765c0e2b65d961f38b134ae5dccb598541ea7d7d7"} Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.155968 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerStarted","Data":"3ca87fd4ed078e2f41e570b6f7db7172077ebc0a5074418fb267bb5d672417ad"} Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.188131 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z45sc" podStartSLOduration=3.258947282 podStartE2EDuration="6.188105364s" podCreationTimestamp="2025-10-01 16:00:14 +0000 UTC" firstStartedPulling="2025-10-01 16:00:16.116194316 +0000 UTC m=+805.466834278" lastFinishedPulling="2025-10-01 16:00:19.045352358 +0000 UTC m=+808.395992360" observedRunningTime="2025-10-01 16:00:20.18144882 +0000 UTC m=+809.532088862" watchObservedRunningTime="2025-10-01 16:00:20.188105364 +0000 UTC m=+809.538745366" Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.210412 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" podStartSLOduration=3.99594802 podStartE2EDuration="4.210394151s" podCreationTimestamp="2025-10-01 16:00:16 +0000 UTC" firstStartedPulling="2025-10-01 16:00:18.130232309 +0000 UTC m=+807.480872271" lastFinishedPulling="2025-10-01 16:00:18.34467844 +0000 UTC m=+807.695318402" observedRunningTime="2025-10-01 16:00:20.208761874 +0000 UTC m=+809.559401846" watchObservedRunningTime="2025-10-01 16:00:20.210394151 +0000 UTC m=+809.561034133" Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.445956 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.446066 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:20 crc kubenswrapper[4688]: I1001 16:00:20.518295 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:21 crc kubenswrapper[4688]: I1001 16:00:21.166590 4688 generic.go:334] "Generic (PLEG): container finished" podID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerID="3ca87fd4ed078e2f41e570b6f7db7172077ebc0a5074418fb267bb5d672417ad" exitCode=0 Oct 01 16:00:21 crc kubenswrapper[4688]: I1001 16:00:21.166762 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerDied","Data":"3ca87fd4ed078e2f41e570b6f7db7172077ebc0a5074418fb267bb5d672417ad"} Oct 01 16:00:21 crc kubenswrapper[4688]: I1001 16:00:21.246616 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.464485 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.571643 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util\") pod \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.571729 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle\") pod \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.572561 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tps94\" (UniqueName: \"kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94\") pod \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\" (UID: \"2f9c1b91-3db2-4b7e-8938-829c2c0d761f\") " Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.572935 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle" (OuterVolumeSpecName: "bundle") pod "2f9c1b91-3db2-4b7e-8938-829c2c0d761f" (UID: "2f9c1b91-3db2-4b7e-8938-829c2c0d761f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.578756 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94" (OuterVolumeSpecName: "kube-api-access-tps94") pod "2f9c1b91-3db2-4b7e-8938-829c2c0d761f" (UID: "2f9c1b91-3db2-4b7e-8938-829c2c0d761f"). InnerVolumeSpecName "kube-api-access-tps94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.587562 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util" (OuterVolumeSpecName: "util") pod "2f9c1b91-3db2-4b7e-8938-829c2c0d761f" (UID: "2f9c1b91-3db2-4b7e-8938-829c2c0d761f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.674154 4688 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.674185 4688 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:22 crc kubenswrapper[4688]: I1001 16:00:22.674194 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tps94\" (UniqueName: \"kubernetes.io/projected/2f9c1b91-3db2-4b7e-8938-829c2c0d761f-kube-api-access-tps94\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:23 crc kubenswrapper[4688]: I1001 16:00:23.184982 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" event={"ID":"2f9c1b91-3db2-4b7e-8938-829c2c0d761f","Type":"ContainerDied","Data":"732ca1d85fc300087df2583aee1c59abeb4dd73fb7cca08fd4f435b229b3425c"} Oct 01 16:00:23 crc kubenswrapper[4688]: I1001 16:00:23.185040 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg" Oct 01 16:00:23 crc kubenswrapper[4688]: I1001 16:00:23.185045 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="732ca1d85fc300087df2583aee1c59abeb4dd73fb7cca08fd4f435b229b3425c" Oct 01 16:00:24 crc kubenswrapper[4688]: I1001 16:00:24.713712 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:24 crc kubenswrapper[4688]: I1001 16:00:24.714642 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-46vfm" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="registry-server" containerID="cri-o://e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7" gracePeriod=2 Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.167311 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.199423 4688 generic.go:334] "Generic (PLEG): container finished" podID="a4128300-2382-4b97-a080-249362e80d6e" containerID="e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7" exitCode=0 Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.199468 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerDied","Data":"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7"} Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.199498 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-46vfm" event={"ID":"a4128300-2382-4b97-a080-249362e80d6e","Type":"ContainerDied","Data":"4fea486182b2cc94f2c769389539195b87bdf876a3bfdda0b7dcef84f17588e5"} Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.199542 4688 scope.go:117] "RemoveContainer" containerID="e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.199674 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-46vfm" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.221242 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content\") pod \"a4128300-2382-4b97-a080-249362e80d6e\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.221319 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities\") pod \"a4128300-2382-4b97-a080-249362e80d6e\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.221369 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d424m\" (UniqueName: \"kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m\") pod \"a4128300-2382-4b97-a080-249362e80d6e\" (UID: \"a4128300-2382-4b97-a080-249362e80d6e\") " Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.223578 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities" (OuterVolumeSpecName: "utilities") pod "a4128300-2382-4b97-a080-249362e80d6e" (UID: "a4128300-2382-4b97-a080-249362e80d6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.227041 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m" (OuterVolumeSpecName: "kube-api-access-d424m") pod "a4128300-2382-4b97-a080-249362e80d6e" (UID: "a4128300-2382-4b97-a080-249362e80d6e"). InnerVolumeSpecName "kube-api-access-d424m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.231679 4688 scope.go:117] "RemoveContainer" containerID="ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.232423 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.232441 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d424m\" (UniqueName: \"kubernetes.io/projected/a4128300-2382-4b97-a080-249362e80d6e-kube-api-access-d424m\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.256782 4688 scope.go:117] "RemoveContainer" containerID="6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.266587 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.266649 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.276561 4688 scope.go:117] "RemoveContainer" containerID="e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7" Oct 01 16:00:25 crc kubenswrapper[4688]: E1001 16:00:25.277779 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7\": container with ID starting with e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7 not found: ID does not exist" containerID="e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.277881 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7"} err="failed to get container status \"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7\": rpc error: code = NotFound desc = could not find container \"e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7\": container with ID starting with e456218e1c964808b6d6bbfbffab93d0c2a235920e7ea59f604efaddb47de7b7 not found: ID does not exist" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.277968 4688 scope.go:117] "RemoveContainer" containerID="ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c" Oct 01 16:00:25 crc kubenswrapper[4688]: E1001 16:00:25.278352 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c\": container with ID starting with ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c not found: ID does not exist" containerID="ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.278385 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c"} err="failed to get container status \"ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c\": rpc error: code = NotFound desc = could not find container \"ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c\": container with ID starting with ede53f10c0f1f3c14c662cfc299c6074e304f3e01ae85bf2850632ae4ce3044c not found: ID does not exist" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.278403 4688 scope.go:117] "RemoveContainer" containerID="6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e" Oct 01 16:00:25 crc kubenswrapper[4688]: E1001 16:00:25.278607 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e\": container with ID starting with 6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e not found: ID does not exist" containerID="6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.278629 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e"} err="failed to get container status \"6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e\": rpc error: code = NotFound desc = could not find container \"6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e\": container with ID starting with 6070ae5f4228bfad04f53ec9438bd3b81eedd97559f6fc8511d1e8547be4e51e not found: ID does not exist" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.301194 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.308800 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4128300-2382-4b97-a080-249362e80d6e" (UID: "a4128300-2382-4b97-a080-249362e80d6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.334057 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4128300-2382-4b97-a080-249362e80d6e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.519986 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:25 crc kubenswrapper[4688]: I1001 16:00:25.526092 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-46vfm"] Oct 01 16:00:26 crc kubenswrapper[4688]: I1001 16:00:26.264316 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.391893 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4128300-2382-4b97-a080-249362e80d6e" path="/var/lib/kubelet/pods/a4128300-2382-4b97-a080-249362e80d6e/volumes" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.912021 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln"] Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.912511 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="extract" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.912629 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="extract" Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.912723 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="extract-content" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.912792 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="extract-content" Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.912882 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="extract-utilities" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.912957 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="extract-utilities" Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.913030 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="registry-server" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.913104 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="registry-server" Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.913174 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="util" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.913239 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="util" Oct 01 16:00:27 crc kubenswrapper[4688]: E1001 16:00:27.913321 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="pull" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.913398 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="pull" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.913627 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f9c1b91-3db2-4b7e-8938-829c2c0d761f" containerName="extract" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.913733 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4128300-2382-4b97-a080-249362e80d6e" containerName="registry-server" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.914609 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.922633 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zd6gf" Oct 01 16:00:27 crc kubenswrapper[4688]: I1001 16:00:27.975974 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln"] Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.068555 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvs8k\" (UniqueName: \"kubernetes.io/projected/ded465e1-ef18-4321-9771-e3a3dd7a92e8-kube-api-access-cvs8k\") pod \"openstack-operator-controller-operator-6445cc7789-t2zln\" (UID: \"ded465e1-ef18-4321-9771-e3a3dd7a92e8\") " pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.120313 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.121781 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.148669 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.169725 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvs8k\" (UniqueName: \"kubernetes.io/projected/ded465e1-ef18-4321-9771-e3a3dd7a92e8-kube-api-access-cvs8k\") pod \"openstack-operator-controller-operator-6445cc7789-t2zln\" (UID: \"ded465e1-ef18-4321-9771-e3a3dd7a92e8\") " pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.219242 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvs8k\" (UniqueName: \"kubernetes.io/projected/ded465e1-ef18-4321-9771-e3a3dd7a92e8-kube-api-access-cvs8k\") pod \"openstack-operator-controller-operator-6445cc7789-t2zln\" (UID: \"ded465e1-ef18-4321-9771-e3a3dd7a92e8\") " pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.231931 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.270548 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.270672 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.270701 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwg8j\" (UniqueName: \"kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.371808 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.372737 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.372306 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.372800 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwg8j\" (UniqueName: \"kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.373163 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.406009 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwg8j\" (UniqueName: \"kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j\") pod \"certified-operators-d6xtn\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.434009 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.668468 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:28 crc kubenswrapper[4688]: I1001 16:00:28.679814 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln"] Oct 01 16:00:29 crc kubenswrapper[4688]: I1001 16:00:29.229525 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" event={"ID":"ded465e1-ef18-4321-9771-e3a3dd7a92e8","Type":"ContainerStarted","Data":"f21e51144e33a6a82ea5960db1cd3a13f727cacdde62933f503842f8be53e5b6"} Oct 01 16:00:29 crc kubenswrapper[4688]: I1001 16:00:29.235198 4688 generic.go:334] "Generic (PLEG): container finished" podID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerID="77d6f1594449eae60a343e305bed2816ceb4a9b68f04a70659ce1c5de104b451" exitCode=0 Oct 01 16:00:29 crc kubenswrapper[4688]: I1001 16:00:29.235261 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerDied","Data":"77d6f1594449eae60a343e305bed2816ceb4a9b68f04a70659ce1c5de104b451"} Oct 01 16:00:29 crc kubenswrapper[4688]: I1001 16:00:29.235301 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerStarted","Data":"6284186e33c63b258e0110dd6fe0ee998e12dba2f60a46a5fe22ac961b43fd3c"} Oct 01 16:00:30 crc kubenswrapper[4688]: I1001 16:00:30.241697 4688 generic.go:334] "Generic (PLEG): container finished" podID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerID="be85cb1d649916afe07b5b2039498ec86809465182c6e5be599c20a203829923" exitCode=0 Oct 01 16:00:30 crc kubenswrapper[4688]: I1001 16:00:30.241752 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerDied","Data":"be85cb1d649916afe07b5b2039498ec86809465182c6e5be599c20a203829923"} Oct 01 16:00:30 crc kubenswrapper[4688]: I1001 16:00:30.909111 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:30 crc kubenswrapper[4688]: I1001 16:00:30.909613 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z45sc" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="registry-server" containerID="cri-o://b63d98fc79af68824e8387a765c0e2b65d961f38b134ae5dccb598541ea7d7d7" gracePeriod=2 Oct 01 16:00:31 crc kubenswrapper[4688]: I1001 16:00:31.252391 4688 generic.go:334] "Generic (PLEG): container finished" podID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerID="b63d98fc79af68824e8387a765c0e2b65d961f38b134ae5dccb598541ea7d7d7" exitCode=0 Oct 01 16:00:31 crc kubenswrapper[4688]: I1001 16:00:31.252478 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerDied","Data":"b63d98fc79af68824e8387a765c0e2b65d961f38b134ae5dccb598541ea7d7d7"} Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.282360 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z45sc" event={"ID":"3639ea8e-a7ca-4827-bd5b-dcf659478c37","Type":"ContainerDied","Data":"249c84b869597692a53e68ad62386d1cd37329ce6ca9c03093862ef0d8f400c3"} Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.282846 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249c84b869597692a53e68ad62386d1cd37329ce6ca9c03093862ef0d8f400c3" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.289077 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.342024 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content\") pod \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.342122 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities\") pod \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.342148 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq9xs\" (UniqueName: \"kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs\") pod \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\" (UID: \"3639ea8e-a7ca-4827-bd5b-dcf659478c37\") " Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.353132 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities" (OuterVolumeSpecName: "utilities") pod "3639ea8e-a7ca-4827-bd5b-dcf659478c37" (UID: "3639ea8e-a7ca-4827-bd5b-dcf659478c37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.358322 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3639ea8e-a7ca-4827-bd5b-dcf659478c37" (UID: "3639ea8e-a7ca-4827-bd5b-dcf659478c37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.370675 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs" (OuterVolumeSpecName: "kube-api-access-wq9xs") pod "3639ea8e-a7ca-4827-bd5b-dcf659478c37" (UID: "3639ea8e-a7ca-4827-bd5b-dcf659478c37"). InnerVolumeSpecName "kube-api-access-wq9xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.444191 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.444234 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3639ea8e-a7ca-4827-bd5b-dcf659478c37-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:33 crc kubenswrapper[4688]: I1001 16:00:33.444243 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq9xs\" (UniqueName: \"kubernetes.io/projected/3639ea8e-a7ca-4827-bd5b-dcf659478c37-kube-api-access-wq9xs\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.290159 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" event={"ID":"ded465e1-ef18-4321-9771-e3a3dd7a92e8","Type":"ContainerStarted","Data":"8113beb45fb95f3df6ebdcd021340ff28cc3419d73e87d7a3d2050b99530d604"} Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.291870 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerStarted","Data":"ccca922293049dd7a1ae48d0b6e1d365470e45bfc04ad7423797e5b32f840342"} Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.291879 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z45sc" Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.319924 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d6xtn" podStartSLOduration=1.929328956 podStartE2EDuration="6.319903311s" podCreationTimestamp="2025-10-01 16:00:28 +0000 UTC" firstStartedPulling="2025-10-01 16:00:29.236834624 +0000 UTC m=+818.587474576" lastFinishedPulling="2025-10-01 16:00:33.627408949 +0000 UTC m=+822.978048931" observedRunningTime="2025-10-01 16:00:34.313399232 +0000 UTC m=+823.664039204" watchObservedRunningTime="2025-10-01 16:00:34.319903311 +0000 UTC m=+823.670543293" Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.327109 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:34 crc kubenswrapper[4688]: I1001 16:00:34.330360 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z45sc"] Oct 01 16:00:35 crc kubenswrapper[4688]: I1001 16:00:35.391863 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" path="/var/lib/kubelet/pods/3639ea8e-a7ca-4827-bd5b-dcf659478c37/volumes" Oct 01 16:00:37 crc kubenswrapper[4688]: I1001 16:00:37.312464 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" event={"ID":"ded465e1-ef18-4321-9771-e3a3dd7a92e8","Type":"ContainerStarted","Data":"ed366e955dce025f226b8862b7205dbe9dbe53b31a125a1cdae66882094d1151"} Oct 01 16:00:37 crc kubenswrapper[4688]: I1001 16:00:37.313387 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:37 crc kubenswrapper[4688]: I1001 16:00:37.353583 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" podStartSLOduration=2.415019911 podStartE2EDuration="10.353568s" podCreationTimestamp="2025-10-01 16:00:27 +0000 UTC" firstStartedPulling="2025-10-01 16:00:28.708510222 +0000 UTC m=+818.059150184" lastFinishedPulling="2025-10-01 16:00:36.647058311 +0000 UTC m=+825.997698273" observedRunningTime="2025-10-01 16:00:37.351584942 +0000 UTC m=+826.702224904" watchObservedRunningTime="2025-10-01 16:00:37.353568 +0000 UTC m=+826.704207962" Oct 01 16:00:38 crc kubenswrapper[4688]: I1001 16:00:38.237316 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6445cc7789-t2zln" Oct 01 16:00:38 crc kubenswrapper[4688]: I1001 16:00:38.445707 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:38 crc kubenswrapper[4688]: I1001 16:00:38.445764 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:38 crc kubenswrapper[4688]: I1001 16:00:38.491265 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:39 crc kubenswrapper[4688]: I1001 16:00:39.379198 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:41 crc kubenswrapper[4688]: I1001 16:00:41.708347 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:41 crc kubenswrapper[4688]: I1001 16:00:41.708603 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d6xtn" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="registry-server" containerID="cri-o://ccca922293049dd7a1ae48d0b6e1d365470e45bfc04ad7423797e5b32f840342" gracePeriod=2 Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.364683 4688 generic.go:334] "Generic (PLEG): container finished" podID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerID="ccca922293049dd7a1ae48d0b6e1d365470e45bfc04ad7423797e5b32f840342" exitCode=0 Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.364920 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerDied","Data":"ccca922293049dd7a1ae48d0b6e1d365470e45bfc04ad7423797e5b32f840342"} Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.790393 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.871780 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities\") pod \"0e5478cd-1a66-424a-a5a6-e90a80125563\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.871867 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwg8j\" (UniqueName: \"kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j\") pod \"0e5478cd-1a66-424a-a5a6-e90a80125563\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.871960 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content\") pod \"0e5478cd-1a66-424a-a5a6-e90a80125563\" (UID: \"0e5478cd-1a66-424a-a5a6-e90a80125563\") " Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.873320 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities" (OuterVolumeSpecName: "utilities") pod "0e5478cd-1a66-424a-a5a6-e90a80125563" (UID: "0e5478cd-1a66-424a-a5a6-e90a80125563"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.884753 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j" (OuterVolumeSpecName: "kube-api-access-qwg8j") pod "0e5478cd-1a66-424a-a5a6-e90a80125563" (UID: "0e5478cd-1a66-424a-a5a6-e90a80125563"). InnerVolumeSpecName "kube-api-access-qwg8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.922451 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e5478cd-1a66-424a-a5a6-e90a80125563" (UID: "0e5478cd-1a66-424a-a5a6-e90a80125563"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.973693 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.973728 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e5478cd-1a66-424a-a5a6-e90a80125563-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:42 crc kubenswrapper[4688]: I1001 16:00:42.973742 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwg8j\" (UniqueName: \"kubernetes.io/projected/0e5478cd-1a66-424a-a5a6-e90a80125563-kube-api-access-qwg8j\") on node \"crc\" DevicePath \"\"" Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.371208 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6xtn" event={"ID":"0e5478cd-1a66-424a-a5a6-e90a80125563","Type":"ContainerDied","Data":"6284186e33c63b258e0110dd6fe0ee998e12dba2f60a46a5fe22ac961b43fd3c"} Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.371267 4688 scope.go:117] "RemoveContainer" containerID="ccca922293049dd7a1ae48d0b6e1d365470e45bfc04ad7423797e5b32f840342" Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.371225 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6xtn" Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.401267 4688 scope.go:117] "RemoveContainer" containerID="be85cb1d649916afe07b5b2039498ec86809465182c6e5be599c20a203829923" Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.422166 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.434213 4688 scope.go:117] "RemoveContainer" containerID="77d6f1594449eae60a343e305bed2816ceb4a9b68f04a70659ce1c5de104b451" Oct 01 16:00:43 crc kubenswrapper[4688]: I1001 16:00:43.441064 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d6xtn"] Oct 01 16:00:45 crc kubenswrapper[4688]: I1001 16:00:45.392200 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" path="/var/lib/kubelet/pods/0e5478cd-1a66-424a-a5a6-e90a80125563/volumes" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.287102 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4"] Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.288883 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.288995 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.289077 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="extract-utilities" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289148 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="extract-utilities" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.289225 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="extract-content" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289295 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="extract-content" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.289356 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289415 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.289483 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="extract-utilities" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289563 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="extract-utilities" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.289662 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="extract-content" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289734 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="extract-content" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.289931 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3639ea8e-a7ca-4827-bd5b-dcf659478c37" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.290375 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e5478cd-1a66-424a-a5a6-e90a80125563" containerName="registry-server" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.291285 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.297963 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6tjvh" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.350581 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.351767 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqfdk\" (UniqueName: \"kubernetes.io/projected/06f11b98-29b6-41f9-a9fa-771b2f9565ad-kube-api-access-fqfdk\") pod \"barbican-operator-controller-manager-6ff8b75857-65bw4\" (UID: \"06f11b98-29b6-41f9-a9fa-771b2f9565ad\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.352083 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.357277 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zjnz5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.357378 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.369695 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.369788 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.377078 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dgx5d" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.389648 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.396926 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.396982 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-548l2"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.397962 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.410540 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-qd88h" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.415050 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.416102 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.418982 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-zzt2h" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.426227 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.427315 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.432303 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-qhhkt" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.452604 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl49b\" (UniqueName: \"kubernetes.io/projected/5fce237a-66be-49ef-90a8-cd0913daf883-kube-api-access-fl49b\") pod \"designate-operator-controller-manager-84f4f7b77b-8pnfq\" (UID: \"5fce237a-66be-49ef-90a8-cd0913daf883\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.452748 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ckdx\" (UniqueName: \"kubernetes.io/projected/09bbe1ee-12b6-4342-b585-f5dfebdcfb79-kube-api-access-6ckdx\") pod \"glance-operator-controller-manager-84958c4d49-548l2\" (UID: \"09bbe1ee-12b6-4342-b585-f5dfebdcfb79\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.452876 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4g2h\" (UniqueName: \"kubernetes.io/projected/5c1fc770-71c1-4e90-b803-d0d020cf22dc-kube-api-access-g4g2h\") pod \"cinder-operator-controller-manager-644bddb6d8-gbg2c\" (UID: \"5c1fc770-71c1-4e90-b803-d0d020cf22dc\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.452997 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqfdk\" (UniqueName: \"kubernetes.io/projected/06f11b98-29b6-41f9-a9fa-771b2f9565ad-kube-api-access-fqfdk\") pod \"barbican-operator-controller-manager-6ff8b75857-65bw4\" (UID: \"06f11b98-29b6-41f9-a9fa-771b2f9565ad\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.454402 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-548l2"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.481997 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.488619 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.506158 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.507168 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.510192 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.510393 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-xv6j9" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.511174 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqfdk\" (UniqueName: \"kubernetes.io/projected/06f11b98-29b6-41f9-a9fa-771b2f9565ad-kube-api-access-fqfdk\") pod \"barbican-operator-controller-manager-6ff8b75857-65bw4\" (UID: \"06f11b98-29b6-41f9-a9fa-771b2f9565ad\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.521552 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.533836 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.534810 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.544596 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-fxnbw" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.551820 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.554484 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.554538 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4g2h\" (UniqueName: \"kubernetes.io/projected/5c1fc770-71c1-4e90-b803-d0d020cf22dc-kube-api-access-g4g2h\") pod \"cinder-operator-controller-manager-644bddb6d8-gbg2c\" (UID: \"5c1fc770-71c1-4e90-b803-d0d020cf22dc\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.555672 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl49b\" (UniqueName: \"kubernetes.io/projected/5fce237a-66be-49ef-90a8-cd0913daf883-kube-api-access-fl49b\") pod \"designate-operator-controller-manager-84f4f7b77b-8pnfq\" (UID: \"5fce237a-66be-49ef-90a8-cd0913daf883\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.555704 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7jt7\" (UniqueName: \"kubernetes.io/projected/e983d733-09fb-4a40-902b-f180b58d5b7b-kube-api-access-q7jt7\") pod \"heat-operator-controller-manager-5d889d78cf-bh7tx\" (UID: \"e983d733-09fb-4a40-902b-f180b58d5b7b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.555726 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln9hr\" (UniqueName: \"kubernetes.io/projected/dc2fb11b-b04c-444c-8623-43a6b566cf6c-kube-api-access-ln9hr\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.555745 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ckdx\" (UniqueName: \"kubernetes.io/projected/09bbe1ee-12b6-4342-b585-f5dfebdcfb79-kube-api-access-6ckdx\") pod \"glance-operator-controller-manager-84958c4d49-548l2\" (UID: \"09bbe1ee-12b6-4342-b585-f5dfebdcfb79\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.555767 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbzxz\" (UniqueName: \"kubernetes.io/projected/44ef9a6f-0918-43c7-b153-abd8c7635131-kube-api-access-xbzxz\") pod \"horizon-operator-controller-manager-9f4696d94-5qm6b\" (UID: \"44ef9a6f-0918-43c7-b153-abd8c7635131\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.569124 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.570039 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.570330 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.570761 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.582213 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-gsmzn" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.588147 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.588999 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.591675 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-zdkcp" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.593056 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-72m9m" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.595241 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.596090 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.612827 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-pp4jl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.612870 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.613070 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.613183 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.625297 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ckdx\" (UniqueName: \"kubernetes.io/projected/09bbe1ee-12b6-4342-b585-f5dfebdcfb79-kube-api-access-6ckdx\") pod \"glance-operator-controller-manager-84958c4d49-548l2\" (UID: \"09bbe1ee-12b6-4342-b585-f5dfebdcfb79\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.663882 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.663922 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twxnx\" (UniqueName: \"kubernetes.io/projected/45f48342-2b0d-4376-87ac-0c775c863326-kube-api-access-twxnx\") pod \"manila-operator-controller-manager-6d68dbc695-mbsl5\" (UID: \"45f48342-2b0d-4376-87ac-0c775c863326\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.663951 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djplf\" (UniqueName: \"kubernetes.io/projected/8fee523c-d09e-422d-94ae-b85585856d27-kube-api-access-djplf\") pod \"mariadb-operator-controller-manager-88c7-x8lrd\" (UID: \"8fee523c-d09e-422d-94ae-b85585856d27\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.663999 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5zsq\" (UniqueName: \"kubernetes.io/projected/57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c-kube-api-access-k5zsq\") pod \"keystone-operator-controller-manager-85595f4675-l62s8\" (UID: \"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c\") " pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.664028 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7jt7\" (UniqueName: \"kubernetes.io/projected/e983d733-09fb-4a40-902b-f180b58d5b7b-kube-api-access-q7jt7\") pod \"heat-operator-controller-manager-5d889d78cf-bh7tx\" (UID: \"e983d733-09fb-4a40-902b-f180b58d5b7b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.664028 4688 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.664064 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln9hr\" (UniqueName: \"kubernetes.io/projected/dc2fb11b-b04c-444c-8623-43a6b566cf6c-kube-api-access-ln9hr\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.664107 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c62ws\" (UniqueName: \"kubernetes.io/projected/6cfc7f51-a252-4b73-954c-b9bb611bf88e-kube-api-access-c62ws\") pod \"neutron-operator-controller-manager-849d5b9b84-h7xkj\" (UID: \"6cfc7f51-a252-4b73-954c-b9bb611bf88e\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.664133 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbzxz\" (UniqueName: \"kubernetes.io/projected/44ef9a6f-0918-43c7-b153-abd8c7635131-kube-api-access-xbzxz\") pod \"horizon-operator-controller-manager-9f4696d94-5qm6b\" (UID: \"44ef9a6f-0918-43c7-b153-abd8c7635131\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.664170 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69cvx\" (UniqueName: \"kubernetes.io/projected/2fbdd013-920f-4e46-8519-3f8bbe24f25d-kube-api-access-69cvx\") pod \"ironic-operator-controller-manager-5cd4858477-b8dfc\" (UID: \"2fbdd013-920f-4e46-8519-3f8bbe24f25d\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:14 crc kubenswrapper[4688]: E1001 16:01:14.664350 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert podName:dc2fb11b-b04c-444c-8623-43a6b566cf6c nodeName:}" failed. No retries permitted until 2025-10-01 16:01:15.164335774 +0000 UTC m=+864.514975736 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert") pod "infra-operator-controller-manager-9d6c5db85-6ccdl" (UID: "dc2fb11b-b04c-444c-8623-43a6b566cf6c") : secret "infra-operator-webhook-server-cert" not found Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.665889 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl49b\" (UniqueName: \"kubernetes.io/projected/5fce237a-66be-49ef-90a8-cd0913daf883-kube-api-access-fl49b\") pod \"designate-operator-controller-manager-84f4f7b77b-8pnfq\" (UID: \"5fce237a-66be-49ef-90a8-cd0913daf883\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.684990 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.685100 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.718129 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.718489 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.719493 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.722213 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4g2h\" (UniqueName: \"kubernetes.io/projected/5c1fc770-71c1-4e90-b803-d0d020cf22dc-kube-api-access-g4g2h\") pod \"cinder-operator-controller-manager-644bddb6d8-gbg2c\" (UID: \"5c1fc770-71c1-4e90-b803-d0d020cf22dc\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.733584 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbzxz\" (UniqueName: \"kubernetes.io/projected/44ef9a6f-0918-43c7-b153-abd8c7635131-kube-api-access-xbzxz\") pod \"horizon-operator-controller-manager-9f4696d94-5qm6b\" (UID: \"44ef9a6f-0918-43c7-b153-abd8c7635131\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.734105 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7jt7\" (UniqueName: \"kubernetes.io/projected/e983d733-09fb-4a40-902b-f180b58d5b7b-kube-api-access-q7jt7\") pod \"heat-operator-controller-manager-5d889d78cf-bh7tx\" (UID: \"e983d733-09fb-4a40-902b-f180b58d5b7b\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.734467 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vzlhj" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.739192 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln9hr\" (UniqueName: \"kubernetes.io/projected/dc2fb11b-b04c-444c-8623-43a6b566cf6c-kube-api-access-ln9hr\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.739441 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.746571 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.747490 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.760226 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.760406 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-628jt" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765070 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69cvx\" (UniqueName: \"kubernetes.io/projected/2fbdd013-920f-4e46-8519-3f8bbe24f25d-kube-api-access-69cvx\") pod \"ironic-operator-controller-manager-5cd4858477-b8dfc\" (UID: \"2fbdd013-920f-4e46-8519-3f8bbe24f25d\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765117 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rk7\" (UniqueName: \"kubernetes.io/projected/3c33683d-9791-4d98-8d6e-0e58f48b4d94-kube-api-access-j8rk7\") pod \"nova-operator-controller-manager-64cd67b5cb-9ppnb\" (UID: \"3c33683d-9791-4d98-8d6e-0e58f48b4d94\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765186 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twxnx\" (UniqueName: \"kubernetes.io/projected/45f48342-2b0d-4376-87ac-0c775c863326-kube-api-access-twxnx\") pod \"manila-operator-controller-manager-6d68dbc695-mbsl5\" (UID: \"45f48342-2b0d-4376-87ac-0c775c863326\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765216 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djplf\" (UniqueName: \"kubernetes.io/projected/8fee523c-d09e-422d-94ae-b85585856d27-kube-api-access-djplf\") pod \"mariadb-operator-controller-manager-88c7-x8lrd\" (UID: \"8fee523c-d09e-422d-94ae-b85585856d27\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765272 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5zsq\" (UniqueName: \"kubernetes.io/projected/57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c-kube-api-access-k5zsq\") pod \"keystone-operator-controller-manager-85595f4675-l62s8\" (UID: \"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c\") " pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.765312 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c62ws\" (UniqueName: \"kubernetes.io/projected/6cfc7f51-a252-4b73-954c-b9bb611bf88e-kube-api-access-c62ws\") pod \"neutron-operator-controller-manager-849d5b9b84-h7xkj\" (UID: \"6cfc7f51-a252-4b73-954c-b9bb611bf88e\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.784087 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.784274 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.845063 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djplf\" (UniqueName: \"kubernetes.io/projected/8fee523c-d09e-422d-94ae-b85585856d27-kube-api-access-djplf\") pod \"mariadb-operator-controller-manager-88c7-x8lrd\" (UID: \"8fee523c-d09e-422d-94ae-b85585856d27\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.858835 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c62ws\" (UniqueName: \"kubernetes.io/projected/6cfc7f51-a252-4b73-954c-b9bb611bf88e-kube-api-access-c62ws\") pod \"neutron-operator-controller-manager-849d5b9b84-h7xkj\" (UID: \"6cfc7f51-a252-4b73-954c-b9bb611bf88e\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.865078 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69cvx\" (UniqueName: \"kubernetes.io/projected/2fbdd013-920f-4e46-8519-3f8bbe24f25d-kube-api-access-69cvx\") pod \"ironic-operator-controller-manager-5cd4858477-b8dfc\" (UID: \"2fbdd013-920f-4e46-8519-3f8bbe24f25d\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.867299 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.876555 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rk7\" (UniqueName: \"kubernetes.io/projected/3c33683d-9791-4d98-8d6e-0e58f48b4d94-kube-api-access-j8rk7\") pod \"nova-operator-controller-manager-64cd67b5cb-9ppnb\" (UID: \"3c33683d-9791-4d98-8d6e-0e58f48b4d94\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.876675 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx8f2\" (UniqueName: \"kubernetes.io/projected/8c13457f-36f6-49b5-9ac2-95298f76cd07-kube-api-access-cx8f2\") pod \"octavia-operator-controller-manager-7b787867f4-rvtpg\" (UID: \"8c13457f-36f6-49b5-9ac2-95298f76cd07\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.895798 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5zsq\" (UniqueName: \"kubernetes.io/projected/57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c-kube-api-access-k5zsq\") pod \"keystone-operator-controller-manager-85595f4675-l62s8\" (UID: \"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c\") " pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.903608 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.915397 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.921193 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twxnx\" (UniqueName: \"kubernetes.io/projected/45f48342-2b0d-4376-87ac-0c775c863326-kube-api-access-twxnx\") pod \"manila-operator-controller-manager-6d68dbc695-mbsl5\" (UID: \"45f48342-2b0d-4376-87ac-0c775c863326\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.923058 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rk7\" (UniqueName: \"kubernetes.io/projected/3c33683d-9791-4d98-8d6e-0e58f48b4d94-kube-api-access-j8rk7\") pod \"nova-operator-controller-manager-64cd67b5cb-9ppnb\" (UID: \"3c33683d-9791-4d98-8d6e-0e58f48b4d94\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.927952 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.935166 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.948867 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.949305 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-jw5gv" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.954946 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.981084 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.981134 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn98g\" (UniqueName: \"kubernetes.io/projected/eb295639-6a81-4c5c-a02b-4101bf064788-kube-api-access-qn98g\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.981224 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx8f2\" (UniqueName: \"kubernetes.io/projected/8c13457f-36f6-49b5-9ac2-95298f76cd07-kube-api-access-cx8f2\") pod \"octavia-operator-controller-manager-7b787867f4-rvtpg\" (UID: \"8c13457f-36f6-49b5-9ac2-95298f76cd07\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.981244 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.985382 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g6725"] Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.993796 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.994373 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-4rwx2" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.996177 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5vf6w" Oct 01 16:01:14 crc kubenswrapper[4688]: I1001 16:01:14.996657 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.029767 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.035051 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx8f2\" (UniqueName: \"kubernetes.io/projected/8c13457f-36f6-49b5-9ac2-95298f76cd07-kube-api-access-cx8f2\") pod \"octavia-operator-controller-manager-7b787867f4-rvtpg\" (UID: \"8c13457f-36f6-49b5-9ac2-95298f76cd07\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.041072 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.057631 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g6725"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.082581 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.083937 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.084046 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.085099 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.085166 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnk44\" (UniqueName: \"kubernetes.io/projected/7c87144a-f596-49d5-8344-b392d2bc4f51-kube-api-access-rnk44\") pod \"placement-operator-controller-manager-589c58c6c-g6725\" (UID: \"7c87144a-f596-49d5-8344-b392d2bc4f51\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.085219 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn98g\" (UniqueName: \"kubernetes.io/projected/eb295639-6a81-4c5c-a02b-4101bf064788-kube-api-access-qn98g\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.085271 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9vmn\" (UniqueName: \"kubernetes.io/projected/fc3e31c2-228a-4ba3-aba1-113ef34d4917-kube-api-access-w9vmn\") pod \"ovn-operator-controller-manager-9976ff44c-tp7l5\" (UID: \"fc3e31c2-228a-4ba3-aba1-113ef34d4917\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.085396 4688 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.085435 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert podName:eb295639-6a81-4c5c-a02b-4101bf064788 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:15.58542157 +0000 UTC m=+864.936061532 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert") pod "openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" (UID: "eb295639-6a81-4c5c-a02b-4101bf064788") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.086064 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.100151 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-cxjj2" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.101060 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.112467 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.113588 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.115829 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.122858 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-npqmn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.132337 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.142577 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-7blrb"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.143583 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.146946 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2ffr2" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.151678 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn98g\" (UniqueName: \"kubernetes.io/projected/eb295639-6a81-4c5c-a02b-4101bf064788-kube-api-access-qn98g\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.161904 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.163464 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.164844 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.167329 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-7blrb"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.169617 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mzzr7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.175459 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187252 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kc2z\" (UniqueName: \"kubernetes.io/projected/a9b72717-b2f8-4340-81ed-88d2c3e0f2d6-kube-api-access-9kc2z\") pod \"telemetry-operator-controller-manager-b8d54b5d7-44nm7\" (UID: \"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187297 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnvr\" (UniqueName: \"kubernetes.io/projected/604860e3-3d42-47b7-9387-c54faa3c8ec7-kube-api-access-grnvr\") pod \"swift-operator-controller-manager-84d6b4b759-f4v99\" (UID: \"604860e3-3d42-47b7-9387-c54faa3c8ec7\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187329 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9vmn\" (UniqueName: \"kubernetes.io/projected/fc3e31c2-228a-4ba3-aba1-113ef34d4917-kube-api-access-w9vmn\") pod \"ovn-operator-controller-manager-9976ff44c-tp7l5\" (UID: \"fc3e31c2-228a-4ba3-aba1-113ef34d4917\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187409 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnk44\" (UniqueName: \"kubernetes.io/projected/7c87144a-f596-49d5-8344-b392d2bc4f51-kube-api-access-rnk44\") pod \"placement-operator-controller-manager-589c58c6c-g6725\" (UID: \"7c87144a-f596-49d5-8344-b392d2bc4f51\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187429 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.187473 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tv7b\" (UniqueName: \"kubernetes.io/projected/733cc8e0-867c-4c7a-b535-c2493a08ed9f-kube-api-access-6tv7b\") pod \"test-operator-controller-manager-85777745bb-7blrb\" (UID: \"733cc8e0-867c-4c7a-b535-c2493a08ed9f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.188321 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.213299 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc2fb11b-b04c-444c-8623-43a6b566cf6c-cert\") pod \"infra-operator-controller-manager-9d6c5db85-6ccdl\" (UID: \"dc2fb11b-b04c-444c-8623-43a6b566cf6c\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.227727 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.228833 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.231483 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4jmgf" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.231870 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.236955 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnk44\" (UniqueName: \"kubernetes.io/projected/7c87144a-f596-49d5-8344-b392d2bc4f51-kube-api-access-rnk44\") pod \"placement-operator-controller-manager-589c58c6c-g6725\" (UID: \"7c87144a-f596-49d5-8344-b392d2bc4f51\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.248987 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9vmn\" (UniqueName: \"kubernetes.io/projected/fc3e31c2-228a-4ba3-aba1-113ef34d4917-kube-api-access-w9vmn\") pod \"ovn-operator-controller-manager-9976ff44c-tp7l5\" (UID: \"fc3e31c2-228a-4ba3-aba1-113ef34d4917\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.253482 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.259704 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.277460 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.278279 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.283750 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-z2msv" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.284763 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.292449 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnvr\" (UniqueName: \"kubernetes.io/projected/604860e3-3d42-47b7-9387-c54faa3c8ec7-kube-api-access-grnvr\") pod \"swift-operator-controller-manager-84d6b4b759-f4v99\" (UID: \"604860e3-3d42-47b7-9387-c54faa3c8ec7\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.293734 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.293880 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fgc\" (UniqueName: \"kubernetes.io/projected/7b66b48a-f1b3-4c89-a879-a662dce03612-kube-api-access-h8fgc\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.293969 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt5gj\" (UniqueName: \"kubernetes.io/projected/52bede76-5e78-4a87-b1ed-4ec996866099-kube-api-access-jt5gj\") pod \"watcher-operator-controller-manager-6b9957f54f-27q6x\" (UID: \"52bede76-5e78-4a87-b1ed-4ec996866099\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.294046 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tv7b\" (UniqueName: \"kubernetes.io/projected/733cc8e0-867c-4c7a-b535-c2493a08ed9f-kube-api-access-6tv7b\") pod \"test-operator-controller-manager-85777745bb-7blrb\" (UID: \"733cc8e0-867c-4c7a-b535-c2493a08ed9f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.294174 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kc2z\" (UniqueName: \"kubernetes.io/projected/a9b72717-b2f8-4340-81ed-88d2c3e0f2d6-kube-api-access-9kc2z\") pod \"telemetry-operator-controller-manager-b8d54b5d7-44nm7\" (UID: \"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.313682 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kc2z\" (UniqueName: \"kubernetes.io/projected/a9b72717-b2f8-4340-81ed-88d2c3e0f2d6-kube-api-access-9kc2z\") pod \"telemetry-operator-controller-manager-b8d54b5d7-44nm7\" (UID: \"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.319720 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.322610 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tv7b\" (UniqueName: \"kubernetes.io/projected/733cc8e0-867c-4c7a-b535-c2493a08ed9f-kube-api-access-6tv7b\") pod \"test-operator-controller-manager-85777745bb-7blrb\" (UID: \"733cc8e0-867c-4c7a-b535-c2493a08ed9f\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.323359 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnvr\" (UniqueName: \"kubernetes.io/projected/604860e3-3d42-47b7-9387-c54faa3c8ec7-kube-api-access-grnvr\") pod \"swift-operator-controller-manager-84d6b4b759-f4v99\" (UID: \"604860e3-3d42-47b7-9387-c54faa3c8ec7\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.344776 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.366967 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.395212 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fgc\" (UniqueName: \"kubernetes.io/projected/7b66b48a-f1b3-4c89-a879-a662dce03612-kube-api-access-h8fgc\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.395254 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4fjd\" (UniqueName: \"kubernetes.io/projected/fa800656-3342-44ac-9ec5-537382b5c67e-kube-api-access-k4fjd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn\" (UID: \"fa800656-3342-44ac-9ec5-537382b5c67e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.395285 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt5gj\" (UniqueName: \"kubernetes.io/projected/52bede76-5e78-4a87-b1ed-4ec996866099-kube-api-access-jt5gj\") pod \"watcher-operator-controller-manager-6b9957f54f-27q6x\" (UID: \"52bede76-5e78-4a87-b1ed-4ec996866099\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.395378 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.395498 4688 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.395559 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert podName:7b66b48a-f1b3-4c89-a879-a662dce03612 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:15.895542217 +0000 UTC m=+865.246182169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert") pod "openstack-operator-controller-manager-5c677785d7-6g6wn" (UID: "7b66b48a-f1b3-4c89-a879-a662dce03612") : secret "webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.433276 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt5gj\" (UniqueName: \"kubernetes.io/projected/52bede76-5e78-4a87-b1ed-4ec996866099-kube-api-access-jt5gj\") pod \"watcher-operator-controller-manager-6b9957f54f-27q6x\" (UID: \"52bede76-5e78-4a87-b1ed-4ec996866099\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.433702 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fgc\" (UniqueName: \"kubernetes.io/projected/7b66b48a-f1b3-4c89-a879-a662dce03612-kube-api-access-h8fgc\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.446307 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.497211 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4fjd\" (UniqueName: \"kubernetes.io/projected/fa800656-3342-44ac-9ec5-537382b5c67e-kube-api-access-k4fjd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn\" (UID: \"fa800656-3342-44ac-9ec5-537382b5c67e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.525550 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4fjd\" (UniqueName: \"kubernetes.io/projected/fa800656-3342-44ac-9ec5-537382b5c67e-kube-api-access-k4fjd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn\" (UID: \"fa800656-3342-44ac-9ec5-537382b5c67e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.577932 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.598115 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.598836 4688 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: E1001 16:01:15.598910 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert podName:eb295639-6a81-4c5c-a02b-4101bf064788 nodeName:}" failed. No retries permitted until 2025-10-01 16:01:16.598893734 +0000 UTC m=+865.949533696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert") pod "openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" (UID: "eb295639-6a81-4c5c-a02b-4101bf064788") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.667796 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.785099 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.911226 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.917702 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7b66b48a-f1b3-4c89-a879-a662dce03612-cert\") pod \"openstack-operator-controller-manager-5c677785d7-6g6wn\" (UID: \"7b66b48a-f1b3-4c89-a879-a662dce03612\") " pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.975759 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.982918 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx"] Oct 01 16:01:15 crc kubenswrapper[4688]: I1001 16:01:15.987705 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4"] Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.000280 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode983d733_09fb_4a40_902b_f180b58d5b7b.slice/crio-93b1bd7dc34e3bca0a8cc4e343d29d68dbd6a734ad098b4bb58fb8d16fd6c945 WatchSource:0}: Error finding container 93b1bd7dc34e3bca0a8cc4e343d29d68dbd6a734ad098b4bb58fb8d16fd6c945: Status 404 returned error can't find the container with id 93b1bd7dc34e3bca0a8cc4e343d29d68dbd6a734ad098b4bb58fb8d16fd6c945 Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.161186 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b"] Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.201419 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c13457f_36f6_49b5_9ac2_95298f76cd07.slice/crio-97e2d34e7af04d1fdf460bd19440dfd0edfc68fc2c5c2cbf010c90048cfef851 WatchSource:0}: Error finding container 97e2d34e7af04d1fdf460bd19440dfd0edfc68fc2c5c2cbf010c90048cfef851: Status 404 returned error can't find the container with id 97e2d34e7af04d1fdf460bd19440dfd0edfc68fc2c5c2cbf010c90048cfef851 Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.202378 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.212695 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.222783 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd"] Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.224273 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fce237a_66be_49ef_90a8_cd0913daf883.slice/crio-f70ed13277fcc2361408840d60ee831f27fe187bc861619a4252295e73299f0d WatchSource:0}: Error finding container f70ed13277fcc2361408840d60ee831f27fe187bc861619a4252295e73299f0d: Status 404 returned error can't find the container with id f70ed13277fcc2361408840d60ee831f27fe187bc861619a4252295e73299f0d Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.227636 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fee523c_d09e_422d_94ae_b85585856d27.slice/crio-e7d6724a03ad23425bf77cdd42f08bcd621273147e259f958e640b91dd5f3ecc WatchSource:0}: Error finding container e7d6724a03ad23425bf77cdd42f08bcd621273147e259f958e640b91dd5f3ecc: Status 404 returned error can't find the container with id e7d6724a03ad23425bf77cdd42f08bcd621273147e259f958e640b91dd5f3ecc Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.228848 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09bbe1ee_12b6_4342_b585_f5dfebdcfb79.slice/crio-796ed7c8e43f08fdcaa761ea01539d59419ce6190ec7237b3a7ff83b409b3d77 WatchSource:0}: Error finding container 796ed7c8e43f08fdcaa761ea01539d59419ce6190ec7237b3a7ff83b409b3d77: Status 404 returned error can't find the container with id 796ed7c8e43f08fdcaa761ea01539d59419ce6190ec7237b3a7ff83b409b3d77 Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.229309 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-548l2"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.361112 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5"] Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.365564 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fbdd013_920f_4e46_8519_3f8bbe24f25d.slice/crio-27f06b1bfaf513c91d857a48cb5c5a3dfa7f5e2994768ed425dba2ae9d3d19d8 WatchSource:0}: Error finding container 27f06b1bfaf513c91d857a48cb5c5a3dfa7f5e2994768ed425dba2ae9d3d19d8: Status 404 returned error can't find the container with id 27f06b1bfaf513c91d857a48cb5c5a3dfa7f5e2994768ed425dba2ae9d3d19d8 Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.374132 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.391602 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.398594 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.401498 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c"] Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.408091 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c1fc770_71c1_4e90_b803_d0d020cf22dc.slice/crio-5d9d99ef52fb447d6702f945760b26c172da3c44c03256bbda4505c42afe6880 WatchSource:0}: Error finding container 5d9d99ef52fb447d6702f945760b26c172da3c44c03256bbda4505c42afe6880: Status 404 returned error can't find the container with id 5d9d99ef52fb447d6702f945760b26c172da3c44c03256bbda4505c42afe6880 Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.551839 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.558426 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-g6725"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.563209 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.613600 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.618684 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.630500 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.636627 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb295639-6a81-4c5c-a02b-4101bf064788-cert\") pod \"openstack-baremetal-operator-controller-manager-77b9676b8cs62rh\" (UID: \"eb295639-6a81-4c5c-a02b-4101bf064788\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.646764 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.653404 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5"] Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.661540 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rnk44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-g6725_openstack-operators(7c87144a-f596-49d5-8344-b392d2bc4f51): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.662070 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j8rk7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-9ppnb_openstack-operators(3c33683d-9791-4d98-8d6e-0e58f48b4d94): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.662659 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.668514 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" event={"ID":"7c87144a-f596-49d5-8344-b392d2bc4f51","Type":"ContainerStarted","Data":"ab770b1f6d2a0e56dbc480541b7695b8a9f6151da97cd55e9ec02e260571dcdd"} Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.670176 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k4fjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn_openstack-operators(fa800656-3342-44ac-9ec5-537382b5c67e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.672061 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" podUID="fa800656-3342-44ac-9ec5-537382b5c67e" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.677088 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" event={"ID":"2fbdd013-920f-4e46-8519-3f8bbe24f25d","Type":"ContainerStarted","Data":"27f06b1bfaf513c91d857a48cb5c5a3dfa7f5e2994768ed425dba2ae9d3d19d8"} Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.677564 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9kc2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-44nm7_openstack-operators(a9b72717-b2f8-4340-81ed-88d2c3e0f2d6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.678379 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jt5gj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6b9957f54f-27q6x_openstack-operators(52bede76-5e78-4a87-b1ed-4ec996866099): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.680799 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" event={"ID":"dc2fb11b-b04c-444c-8623-43a6b566cf6c","Type":"ContainerStarted","Data":"01eea9f3cfa1cd599a3bd516347ee79cd627b47aa791e71b513fa53581085044"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.682593 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" event={"ID":"7b66b48a-f1b3-4c89-a879-a662dce03612","Type":"ContainerStarted","Data":"20f9984d609e98184dee2ce8542a99894aee5da01def4daa2e4669dd2e7ad848"} Oct 01 16:01:16 crc kubenswrapper[4688]: W1001 16:01:16.684355 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc3e31c2_228a_4ba3_aba1_113ef34d4917.slice/crio-f90b2560bc256385792e909c1ecf366ae66a9d0b9f7be45a757795ceaa95980c WatchSource:0}: Error finding container f90b2560bc256385792e909c1ecf366ae66a9d0b9f7be45a757795ceaa95980c: Status 404 returned error can't find the container with id f90b2560bc256385792e909c1ecf366ae66a9d0b9f7be45a757795ceaa95980c Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.684808 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" event={"ID":"06f11b98-29b6-41f9-a9fa-771b2f9565ad","Type":"ContainerStarted","Data":"70085d76af91012931e44115dedc30a59544a4958db3e09fb61a00e43c6e607c"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.686075 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" event={"ID":"8c13457f-36f6-49b5-9ac2-95298f76cd07","Type":"ContainerStarted","Data":"97e2d34e7af04d1fdf460bd19440dfd0edfc68fc2c5c2cbf010c90048cfef851"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.690665 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" event={"ID":"6cfc7f51-a252-4b73-954c-b9bb611bf88e","Type":"ContainerStarted","Data":"311908f2e9fa14798729bd00926ff5dc8a0928cff61fef9ad0bd6021c514d3d2"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.692267 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" event={"ID":"44ef9a6f-0918-43c7-b153-abd8c7635131","Type":"ContainerStarted","Data":"74e5aefc50d6f5315bb38909ab77ddf371d9b8c56d1375b9fd625b31ec05c632"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.693096 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" event={"ID":"09bbe1ee-12b6-4342-b585-f5dfebdcfb79","Type":"ContainerStarted","Data":"796ed7c8e43f08fdcaa761ea01539d59419ce6190ec7237b3a7ff83b409b3d77"} Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.694654 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-grnvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-f4v99_openstack-operators(604860e3-3d42-47b7-9387-c54faa3c8ec7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.694895 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" event={"ID":"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c","Type":"ContainerStarted","Data":"7d7005c7718bc9311c450daea973ad6576f8a2baf5272b8ba4a72a871fbb4474"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.704328 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99"] Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.704360 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-7blrb"] Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.709777 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w9vmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-tp7l5_openstack-operators(fc3e31c2-228a-4ba3-aba1-113ef34d4917): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.710421 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" event={"ID":"8fee523c-d09e-422d-94ae-b85585856d27","Type":"ContainerStarted","Data":"e7d6724a03ad23425bf77cdd42f08bcd621273147e259f958e640b91dd5f3ecc"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.717719 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" event={"ID":"e983d733-09fb-4a40-902b-f180b58d5b7b","Type":"ContainerStarted","Data":"93b1bd7dc34e3bca0a8cc4e343d29d68dbd6a734ad098b4bb58fb8d16fd6c945"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.724234 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" event={"ID":"45f48342-2b0d-4376-87ac-0c775c863326","Type":"ContainerStarted","Data":"c0b505eb2310db9f8488f50df28a5fd635b71cf45510a1f07eb4914876b6157e"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.729679 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" event={"ID":"5fce237a-66be-49ef-90a8-cd0913daf883","Type":"ContainerStarted","Data":"f70ed13277fcc2361408840d60ee831f27fe187bc861619a4252295e73299f0d"} Oct 01 16:01:16 crc kubenswrapper[4688]: E1001 16:01:16.729949 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6tv7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-7blrb_openstack-operators(733cc8e0-867c-4c7a-b535-c2493a08ed9f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.732982 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" event={"ID":"3c33683d-9791-4d98-8d6e-0e58f48b4d94","Type":"ContainerStarted","Data":"e0294be14ab47fd7c51e3104cabc05e5a5ab7cf6fe7eb0e8ff04acc2764d8caa"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.737174 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" event={"ID":"5c1fc770-71c1-4e90-b803-d0d020cf22dc","Type":"ContainerStarted","Data":"5d9d99ef52fb447d6702f945760b26c172da3c44c03256bbda4505c42afe6880"} Oct 01 16:01:16 crc kubenswrapper[4688]: I1001 16:01:16.823799 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.004204 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" podUID="7c87144a-f596-49d5-8344-b392d2bc4f51" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.059379 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" podUID="3c33683d-9791-4d98-8d6e-0e58f48b4d94" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.082827 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" podUID="a9b72717-b2f8-4340-81ed-88d2c3e0f2d6" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.086031 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" podUID="52bede76-5e78-4a87-b1ed-4ec996866099" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.234053 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" podUID="fc3e31c2-228a-4ba3-aba1-113ef34d4917" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.261775 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" podUID="604860e3-3d42-47b7-9387-c54faa3c8ec7" Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.448351 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" podUID="733cc8e0-867c-4c7a-b535-c2493a08ed9f" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.550661 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh"] Oct 01 16:01:17 crc kubenswrapper[4688]: W1001 16:01:17.581534 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb295639_6a81_4c5c_a02b_4101bf064788.slice/crio-41d3cc68521bd3ee013fdaecd1182508b31bf22f0a2bdfc1ec17e9865cc30036 WatchSource:0}: Error finding container 41d3cc68521bd3ee013fdaecd1182508b31bf22f0a2bdfc1ec17e9865cc30036: Status 404 returned error can't find the container with id 41d3cc68521bd3ee013fdaecd1182508b31bf22f0a2bdfc1ec17e9865cc30036 Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.751669 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" event={"ID":"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6","Type":"ContainerStarted","Data":"c59cc8631c53ec6f40a43d2a19efea381657fd283e902b87ba135de0e1ca1713"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.751715 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" event={"ID":"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6","Type":"ContainerStarted","Data":"4ba92e26199f642143b873a6a2c1a96f872f7e28ce7ec7ae77d2c098be06f03a"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.753326 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" podUID="a9b72717-b2f8-4340-81ed-88d2c3e0f2d6" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.755793 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" event={"ID":"7c87144a-f596-49d5-8344-b392d2bc4f51","Type":"ContainerStarted","Data":"b2651bb6a407e642da9d03ba50fc20b0ba09d83b6a211985a49ced174dccd4eb"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.759122 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" podUID="7c87144a-f596-49d5-8344-b392d2bc4f51" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.782308 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" event={"ID":"fc3e31c2-228a-4ba3-aba1-113ef34d4917","Type":"ContainerStarted","Data":"c7b8f7dcc56cc2455e4565e52cd6c0e196671fdb99b455eff194f827700057a0"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.782357 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" event={"ID":"fc3e31c2-228a-4ba3-aba1-113ef34d4917","Type":"ContainerStarted","Data":"f90b2560bc256385792e909c1ecf366ae66a9d0b9f7be45a757795ceaa95980c"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.797061 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" podUID="fc3e31c2-228a-4ba3-aba1-113ef34d4917" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.814989 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" event={"ID":"3c33683d-9791-4d98-8d6e-0e58f48b4d94","Type":"ContainerStarted","Data":"4154fdaae0c42f53aac85ef2df1bae3081c2b0ecca8e9c29a75b337d6c7f9c20"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.815971 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" podUID="3c33683d-9791-4d98-8d6e-0e58f48b4d94" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.830170 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" event={"ID":"604860e3-3d42-47b7-9387-c54faa3c8ec7","Type":"ContainerStarted","Data":"a62dedbddb272ac0f0e806abb111d4624aa239efe79eafb8d5ba64a7214910d1"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.830219 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" event={"ID":"604860e3-3d42-47b7-9387-c54faa3c8ec7","Type":"ContainerStarted","Data":"f7ae1636a17007a863549564ce3e4d767fb30adad485ce9ea6fec623477572d9"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.836607 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" podUID="604860e3-3d42-47b7-9387-c54faa3c8ec7" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.840755 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" event={"ID":"7b66b48a-f1b3-4c89-a879-a662dce03612","Type":"ContainerStarted","Data":"e40a6ff0406f2c70024bcccd9b1d8d4d3bae89715fc36d69c3a1057e58cb4d3b"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.840860 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" event={"ID":"7b66b48a-f1b3-4c89-a879-a662dce03612","Type":"ContainerStarted","Data":"e03a1731f483c57947b960260c73b0e0a7d70a6ec8966da0677eca12b96c0f27"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.841908 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.853024 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" event={"ID":"52bede76-5e78-4a87-b1ed-4ec996866099","Type":"ContainerStarted","Data":"f3ff3f91d38d18ab94e88c568f7f829c28699890682501be2691a1f90659357a"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.853060 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" event={"ID":"52bede76-5e78-4a87-b1ed-4ec996866099","Type":"ContainerStarted","Data":"d51c5fd85bd509ba41eb819de97b4725c3a66c50ede1727e3a8e560b55d0345b"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.855050 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" podUID="52bede76-5e78-4a87-b1ed-4ec996866099" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.860774 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" event={"ID":"fa800656-3342-44ac-9ec5-537382b5c67e","Type":"ContainerStarted","Data":"6a2f7792e3b675f52729e941a42e1060e841814d92878489f3dc6ea1acca7522"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.863391 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" podUID="fa800656-3342-44ac-9ec5-537382b5c67e" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.865542 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" event={"ID":"733cc8e0-867c-4c7a-b535-c2493a08ed9f","Type":"ContainerStarted","Data":"92a4a43d1d6d08ea328a1977cb7523b02c4ac5f5c5ee0b1df0297d1292326fcb"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.865594 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" event={"ID":"733cc8e0-867c-4c7a-b535-c2493a08ed9f","Type":"ContainerStarted","Data":"516ea573b3a3984d2d33ce58f81344ffd5162fe30775c2d9e384f1b9d9717e9e"} Oct 01 16:01:17 crc kubenswrapper[4688]: E1001 16:01:17.867901 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" podUID="733cc8e0-867c-4c7a-b535-c2493a08ed9f" Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.879951 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" event={"ID":"eb295639-6a81-4c5c-a02b-4101bf064788","Type":"ContainerStarted","Data":"41d3cc68521bd3ee013fdaecd1182508b31bf22f0a2bdfc1ec17e9865cc30036"} Oct 01 16:01:17 crc kubenswrapper[4688]: I1001 16:01:17.889132 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" podStartSLOduration=2.889117261 podStartE2EDuration="2.889117261s" podCreationTimestamp="2025-10-01 16:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:01:17.888354469 +0000 UTC m=+867.238994431" watchObservedRunningTime="2025-10-01 16:01:17.889117261 +0000 UTC m=+867.239757223" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905145 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" podUID="3c33683d-9791-4d98-8d6e-0e58f48b4d94" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905194 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" podUID="fc3e31c2-228a-4ba3-aba1-113ef34d4917" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905248 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" podUID="733cc8e0-867c-4c7a-b535-c2493a08ed9f" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905265 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" podUID="604860e3-3d42-47b7-9387-c54faa3c8ec7" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905807 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" podUID="7c87144a-f596-49d5-8344-b392d2bc4f51" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.905881 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" podUID="a9b72717-b2f8-4340-81ed-88d2c3e0f2d6" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.906419 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" podUID="fa800656-3342-44ac-9ec5-537382b5c67e" Oct 01 16:01:18 crc kubenswrapper[4688]: E1001 16:01:18.906553 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" podUID="52bede76-5e78-4a87-b1ed-4ec996866099" Oct 01 16:01:25 crc kubenswrapper[4688]: I1001 16:01:25.052817 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:01:25 crc kubenswrapper[4688]: I1001 16:01:25.053554 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:01:25 crc kubenswrapper[4688]: I1001 16:01:25.980797 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5c677785d7-6g6wn" Oct 01 16:01:30 crc kubenswrapper[4688]: E1001 16:01:30.523896 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397" Oct 01 16:01:30 crc kubenswrapper[4688]: E1001 16:01:30.524360 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xbzxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-9f4696d94-5qm6b_openstack-operators(44ef9a6f-0918-43c7-b153-abd8c7635131): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:31 crc kubenswrapper[4688]: E1001 16:01:31.170238 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72" Oct 01 16:01:31 crc kubenswrapper[4688]: E1001 16:01:31.170416 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6ckdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-84958c4d49-548l2_openstack-operators(09bbe1ee-12b6-4342-b585-f5dfebdcfb79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:31 crc kubenswrapper[4688]: E1001 16:01:31.809904 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:bb39758cc8cd0d2cd02841dc81b53fd88647e2db15ee16cdd8c44d4098a942fd" Oct 01 16:01:31 crc kubenswrapper[4688]: E1001 16:01:31.810035 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:bb39758cc8cd0d2cd02841dc81b53fd88647e2db15ee16cdd8c44d4098a942fd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fqfdk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-6ff8b75857-65bw4_openstack-operators(06f11b98-29b6-41f9-a9fa-771b2f9565ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:34 crc kubenswrapper[4688]: E1001 16:01:34.063092 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2" Oct 01 16:01:34 crc kubenswrapper[4688]: E1001 16:01:34.064073 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fl49b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-84f4f7b77b-8pnfq_openstack-operators(5fce237a-66be-49ef-90a8-cd0913daf883): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:35 crc kubenswrapper[4688]: E1001 16:01:35.480510 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884" Oct 01 16:01:35 crc kubenswrapper[4688]: E1001 16:01:35.481373 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twxnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-mbsl5_openstack-operators(45f48342-2b0d-4376-87ac-0c775c863326): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:35 crc kubenswrapper[4688]: E1001 16:01:35.575018 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.73:5001/openstack-k8s-operators/keystone-operator:062cef17cc20f319c285a7bbdc7f2e5390a9b4d4" Oct 01 16:01:35 crc kubenswrapper[4688]: E1001 16:01:35.575082 4688 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.73:5001/openstack-k8s-operators/keystone-operator:062cef17cc20f319c285a7bbdc7f2e5390a9b4d4" Oct 01 16:01:35 crc kubenswrapper[4688]: E1001 16:01:35.575268 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.73:5001/openstack-k8s-operators/keystone-operator:062cef17cc20f319c285a7bbdc7f2e5390a9b4d4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5zsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-85595f4675-l62s8_openstack-operators(57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:01:46 crc kubenswrapper[4688]: E1001 16:01:46.764868 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" podUID="09bbe1ee-12b6-4342-b585-f5dfebdcfb79" Oct 01 16:01:47 crc kubenswrapper[4688]: E1001 16:01:47.094869 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" podUID="5fce237a-66be-49ef-90a8-cd0913daf883" Oct 01 16:01:47 crc kubenswrapper[4688]: E1001 16:01:47.105121 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" podUID="44ef9a6f-0918-43c7-b153-abd8c7635131" Oct 01 16:01:47 crc kubenswrapper[4688]: E1001 16:01:47.111931 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" podUID="45f48342-2b0d-4376-87ac-0c775c863326" Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.116757 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" event={"ID":"45f48342-2b0d-4376-87ac-0c775c863326","Type":"ContainerStarted","Data":"75d0dfebda22911c2f337375dc8a22a3cefae0c50ab8f27bfe04e17a94aa1059"} Oct 01 16:01:47 crc kubenswrapper[4688]: E1001 16:01:47.134294 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" podUID="57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c" Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.140633 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" event={"ID":"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c","Type":"ContainerStarted","Data":"df12dc6abb6a9dde892f77dbb6143c00bf160c9a308cfaf27a26c662fcfa0446"} Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.152728 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" event={"ID":"5fce237a-66be-49ef-90a8-cd0913daf883","Type":"ContainerStarted","Data":"acbf09c7bdc0bdfdea62cc3e8e871e453228063892836b26c8b12c7ce5f3b644"} Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.171754 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" event={"ID":"8fee523c-d09e-422d-94ae-b85585856d27","Type":"ContainerStarted","Data":"c6902a8fde66b75e8abec37d7011e8055568b4fea1ebd7f91aa810362696f5d5"} Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.184323 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" event={"ID":"44ef9a6f-0918-43c7-b153-abd8c7635131","Type":"ContainerStarted","Data":"486c996874f8b0045b0f4f0a1f177fd0aa1751ef35f3e676790e58be2ce9c94c"} Oct 01 16:01:47 crc kubenswrapper[4688]: I1001 16:01:47.200739 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" event={"ID":"09bbe1ee-12b6-4342-b585-f5dfebdcfb79","Type":"ContainerStarted","Data":"f28bb3c3554ba5234009a70c889f1e9c4cc9f182b28e7433176ca4621ddbfc7a"} Oct 01 16:01:47 crc kubenswrapper[4688]: E1001 16:01:47.527689 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" podUID="06f11b98-29b6-41f9-a9fa-771b2f9565ad" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.220064 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" event={"ID":"06f11b98-29b6-41f9-a9fa-771b2f9565ad","Type":"ContainerStarted","Data":"50e003a182df8ce5def160cb810a1f5a355820639c903fd4e91fd2425e0f561d"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.222793 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" event={"ID":"dc2fb11b-b04c-444c-8623-43a6b566cf6c","Type":"ContainerStarted","Data":"6d9977ec9e746d6e1733f70c70826d191ed5e153ecd1744e7c0e1020250fa292"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.223296 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" event={"ID":"dc2fb11b-b04c-444c-8623-43a6b566cf6c","Type":"ContainerStarted","Data":"20cd44bdd09bca64856f18334f930fdd4266d865f1d8aaf7de9933252cbac95e"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.224049 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.229825 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" event={"ID":"e983d733-09fb-4a40-902b-f180b58d5b7b","Type":"ContainerStarted","Data":"1307e26c3ba7f700a1343e5bfda8f41f0928a4edd5327f316a1e2ace77072aed"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.229876 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" event={"ID":"e983d733-09fb-4a40-902b-f180b58d5b7b","Type":"ContainerStarted","Data":"3a7b3deefa525b1c83f7fc93dfdaac2797eba857706ff86c1df054379353db23"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.229898 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.234004 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" event={"ID":"604860e3-3d42-47b7-9387-c54faa3c8ec7","Type":"ContainerStarted","Data":"6cde4741ad33e871fc6f5083ebdcadaf83dc4af78503fe78d99427092aa51465"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.234677 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.236779 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" event={"ID":"52bede76-5e78-4a87-b1ed-4ec996866099","Type":"ContainerStarted","Data":"7033e434bd1c119e50eb2d75e75b0dd3dadb5ac60aab9dfa596bab13c5f22805"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.237378 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.239085 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" event={"ID":"fa800656-3342-44ac-9ec5-537382b5c67e","Type":"ContainerStarted","Data":"6d9d44113a719c3abec62af7541eef09515d523e51241947448e293f12fb0ab6"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.241480 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" event={"ID":"6cfc7f51-a252-4b73-954c-b9bb611bf88e","Type":"ContainerStarted","Data":"17862d509b1d79992b92f1cd77df93320553204cd76a2f0bf01179474d2f6834"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.241515 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" event={"ID":"6cfc7f51-a252-4b73-954c-b9bb611bf88e","Type":"ContainerStarted","Data":"74c00e0a245ed2796bd112186a683022aa3365a36846887d478f3fe44c077dd1"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.241692 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.249832 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" event={"ID":"2fbdd013-920f-4e46-8519-3f8bbe24f25d","Type":"ContainerStarted","Data":"40aa8989c1c36c3bcc33b0b52ea9d07d69de8ba8b37550453926f342af18233d"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.249884 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" event={"ID":"2fbdd013-920f-4e46-8519-3f8bbe24f25d","Type":"ContainerStarted","Data":"75256de42e687d249fb04a96b9b925773a6fe18c39b8df0a1793ad95bf66be88"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.250679 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.255108 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" event={"ID":"a9b72717-b2f8-4340-81ed-88d2c3e0f2d6","Type":"ContainerStarted","Data":"d0b0d13e7384a73ead1d121a92edec2156611c1457ebd551298daa320fdcf986"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.255905 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.262705 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" event={"ID":"fc3e31c2-228a-4ba3-aba1-113ef34d4917","Type":"ContainerStarted","Data":"5a2bb523e3a1d8adb951f43d8b48871523ddf2d972150b29f78a27386f652109"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.263393 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.264673 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" event={"ID":"8c13457f-36f6-49b5-9ac2-95298f76cd07","Type":"ContainerStarted","Data":"9985d925a4647fabe42d20b9f34ca09ca57b9b2d960bebd34b6bf8eca34adb65"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.264706 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" event={"ID":"8c13457f-36f6-49b5-9ac2-95298f76cd07","Type":"ContainerStarted","Data":"668bb9d5bbad582889d6dded14608a400ba8654af27c3608f442a0f67f34be5e"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.265283 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.267575 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" event={"ID":"8fee523c-d09e-422d-94ae-b85585856d27","Type":"ContainerStarted","Data":"8b0ea49bf5a2d1625195b378a247c9393953917b45492d55b01bbd2218925c17"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.268052 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.270881 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" event={"ID":"3c33683d-9791-4d98-8d6e-0e58f48b4d94","Type":"ContainerStarted","Data":"3e0c7d76d88730845ff8f7b1959765e5c7d2d41b6d924df7f02f2fb713d4e194"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.271385 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.274430 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" event={"ID":"7c87144a-f596-49d5-8344-b392d2bc4f51","Type":"ContainerStarted","Data":"67a2b0a3e9f21c6c2e9dfc192ffd3d3561ad4db4ef63981eb6499d7885fb300e"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.274874 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.308652 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" event={"ID":"733cc8e0-867c-4c7a-b535-c2493a08ed9f","Type":"ContainerStarted","Data":"e7d5d1f62f7063177529f6d347fcb37b73b0750cb110eb0229fd944063950769"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.309426 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.312002 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" event={"ID":"eb295639-6a81-4c5c-a02b-4101bf064788","Type":"ContainerStarted","Data":"d0432ff776d4e0970da2933b356cc7b11998082b1d5e0ad2a8eac23851492539"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.334761 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" event={"ID":"5c1fc770-71c1-4e90-b803-d0d020cf22dc","Type":"ContainerStarted","Data":"61f42a5c03587ca545d3c1d17c02caebc3a8ba8ca5ee1fef6636512327f3b4f4"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.334823 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" event={"ID":"5c1fc770-71c1-4e90-b803-d0d020cf22dc","Type":"ContainerStarted","Data":"0a37dd4bab85082bac76fb07fd1f54f067462ffe63134b7803255a3968210d80"} Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.335063 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.537220 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" podStartSLOduration=4.6784227430000005 podStartE2EDuration="34.537200438s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.661373057 +0000 UTC m=+866.012013019" lastFinishedPulling="2025-10-01 16:01:46.520150742 +0000 UTC m=+895.870790714" observedRunningTime="2025-10-01 16:01:48.530900168 +0000 UTC m=+897.881540130" watchObservedRunningTime="2025-10-01 16:01:48.537200438 +0000 UTC m=+897.887840410" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.539793 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" podStartSLOduration=4.68043077 podStartE2EDuration="34.539779381s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.694552005 +0000 UTC m=+866.045191967" lastFinishedPulling="2025-10-01 16:01:46.553900616 +0000 UTC m=+895.904540578" observedRunningTime="2025-10-01 16:01:48.377910058 +0000 UTC m=+897.728550020" watchObservedRunningTime="2025-10-01 16:01:48.539779381 +0000 UTC m=+897.890419343" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.591548 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" podStartSLOduration=10.816701787 podStartE2EDuration="34.591515819s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.409692129 +0000 UTC m=+865.760332091" lastFinishedPulling="2025-10-01 16:01:40.184506121 +0000 UTC m=+889.535146123" observedRunningTime="2025-10-01 16:01:48.585333732 +0000 UTC m=+897.935973694" watchObservedRunningTime="2025-10-01 16:01:48.591515819 +0000 UTC m=+897.942155771" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.712569 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" podStartSLOduration=4.918486308 podStartE2EDuration="34.712550425s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.66182254 +0000 UTC m=+866.012462502" lastFinishedPulling="2025-10-01 16:01:46.455886657 +0000 UTC m=+895.806526619" observedRunningTime="2025-10-01 16:01:48.683438624 +0000 UTC m=+898.034078586" watchObservedRunningTime="2025-10-01 16:01:48.712550425 +0000 UTC m=+898.063190387" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.774075 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn" podStartSLOduration=3.9803409849999998 podStartE2EDuration="33.774052902s" podCreationTimestamp="2025-10-01 16:01:15 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.670077596 +0000 UTC m=+866.020717548" lastFinishedPulling="2025-10-01 16:01:46.463789503 +0000 UTC m=+895.814429465" observedRunningTime="2025-10-01 16:01:48.758097576 +0000 UTC m=+898.108737538" watchObservedRunningTime="2025-10-01 16:01:48.774052902 +0000 UTC m=+898.124692864" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.809379 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" podStartSLOduration=4.966867691 podStartE2EDuration="34.80936011s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.677667423 +0000 UTC m=+866.028307385" lastFinishedPulling="2025-10-01 16:01:46.520159842 +0000 UTC m=+895.870799804" observedRunningTime="2025-10-01 16:01:48.806421156 +0000 UTC m=+898.157061108" watchObservedRunningTime="2025-10-01 16:01:48.80936011 +0000 UTC m=+898.160000062" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.835364 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" podStartSLOduration=5.080198046 podStartE2EDuration="34.835344952s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.709657586 +0000 UTC m=+866.060297548" lastFinishedPulling="2025-10-01 16:01:46.464804482 +0000 UTC m=+895.815444454" observedRunningTime="2025-10-01 16:01:48.83175572 +0000 UTC m=+898.182395672" watchObservedRunningTime="2025-10-01 16:01:48.835344952 +0000 UTC m=+898.185984914" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.893059 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" podStartSLOduration=5.050219951 podStartE2EDuration="34.8930391s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.677352944 +0000 UTC m=+866.027992906" lastFinishedPulling="2025-10-01 16:01:46.520172083 +0000 UTC m=+895.870812055" observedRunningTime="2025-10-01 16:01:48.86711531 +0000 UTC m=+898.217755292" watchObservedRunningTime="2025-10-01 16:01:48.8930391 +0000 UTC m=+898.243679062" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.923022 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" podStartSLOduration=11.110094856 podStartE2EDuration="34.923001856s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.371478538 +0000 UTC m=+865.722118500" lastFinishedPulling="2025-10-01 16:01:40.184385498 +0000 UTC m=+889.535025500" observedRunningTime="2025-10-01 16:01:48.897627831 +0000 UTC m=+898.248267793" watchObservedRunningTime="2025-10-01 16:01:48.923001856 +0000 UTC m=+898.273641818" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.962868 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" podStartSLOduration=11.177168172 podStartE2EDuration="34.962850284s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.398773918 +0000 UTC m=+865.749413880" lastFinishedPulling="2025-10-01 16:01:40.18445599 +0000 UTC m=+889.535095992" observedRunningTime="2025-10-01 16:01:48.959420396 +0000 UTC m=+898.310060358" watchObservedRunningTime="2025-10-01 16:01:48.962850284 +0000 UTC m=+898.313490246" Oct 01 16:01:48 crc kubenswrapper[4688]: I1001 16:01:48.964824 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" podStartSLOduration=10.986266269 podStartE2EDuration="34.9648169s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.206042583 +0000 UTC m=+865.556682545" lastFinishedPulling="2025-10-01 16:01:40.184593204 +0000 UTC m=+889.535233176" observedRunningTime="2025-10-01 16:01:48.928266076 +0000 UTC m=+898.278906038" watchObservedRunningTime="2025-10-01 16:01:48.9648169 +0000 UTC m=+898.315456862" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.060494 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" podStartSLOduration=10.88331872 podStartE2EDuration="35.060473962s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.007208195 +0000 UTC m=+865.357848157" lastFinishedPulling="2025-10-01 16:01:40.184363407 +0000 UTC m=+889.535003399" observedRunningTime="2025-10-01 16:01:49.009800855 +0000 UTC m=+898.360440817" watchObservedRunningTime="2025-10-01 16:01:49.060473962 +0000 UTC m=+898.411113934" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.061064 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" podStartSLOduration=11.10846325 podStartE2EDuration="35.061058479s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.233443316 +0000 UTC m=+865.584083278" lastFinishedPulling="2025-10-01 16:01:40.186038505 +0000 UTC m=+889.536678507" observedRunningTime="2025-10-01 16:01:49.055745737 +0000 UTC m=+898.406385699" watchObservedRunningTime="2025-10-01 16:01:49.061058479 +0000 UTC m=+898.411698441" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.087971 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" podStartSLOduration=5.274693182 podStartE2EDuration="35.087952967s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.729776551 +0000 UTC m=+866.080416503" lastFinishedPulling="2025-10-01 16:01:46.543036316 +0000 UTC m=+895.893676288" observedRunningTime="2025-10-01 16:01:49.086397632 +0000 UTC m=+898.437037584" watchObservedRunningTime="2025-10-01 16:01:49.087952967 +0000 UTC m=+898.438592929" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.342204 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" event={"ID":"5fce237a-66be-49ef-90a8-cd0913daf883","Type":"ContainerStarted","Data":"58db5005d1e4df92b4a3031a52efc6c54faf7d2743f93d3288a0d32cd7f26821"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.343108 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.344584 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" event={"ID":"eb295639-6a81-4c5c-a02b-4101bf064788","Type":"ContainerStarted","Data":"b8be6b092c7a4ee78956b9a5ce56ab38968303440da0aca7809e5b602b3593f5"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.344923 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.346547 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" event={"ID":"44ef9a6f-0918-43c7-b153-abd8c7635131","Type":"ContainerStarted","Data":"dcb7215b466e1fa53eaa3b88e022022e0eaabe1f833f6ae6fc6dca7997f8dbf8"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.346693 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.348172 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" event={"ID":"09bbe1ee-12b6-4342-b585-f5dfebdcfb79","Type":"ContainerStarted","Data":"ac4433ef6bf0384b4e90046e8076bb0d037b8e1c63f4c30b34f0f19d0447d2b2"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.348209 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.349629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" event={"ID":"45f48342-2b0d-4376-87ac-0c775c863326","Type":"ContainerStarted","Data":"990b668896aceb6300629a08065af862e4e9e119edd15f3e1f2b79bb8bd24610"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.349980 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.351334 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" event={"ID":"06f11b98-29b6-41f9-a9fa-771b2f9565ad","Type":"ContainerStarted","Data":"763d4623ee60cbf8a14b885aeeb381836d2d392ef927b2eacad298f67d5e0d4f"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.351704 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.353726 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" event={"ID":"57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c","Type":"ContainerStarted","Data":"a76f60613389338c3a50c62909413a997cd372256d1c79d004afd1bc22225792"} Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.353865 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.368477 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" podStartSLOduration=3.56207297 podStartE2EDuration="35.368460858s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.226539939 +0000 UTC m=+865.577179901" lastFinishedPulling="2025-10-01 16:01:48.032927827 +0000 UTC m=+897.383567789" observedRunningTime="2025-10-01 16:01:49.36643478 +0000 UTC m=+898.717074742" watchObservedRunningTime="2025-10-01 16:01:49.368460858 +0000 UTC m=+898.719100820" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.369300 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" podStartSLOduration=11.595440298 podStartE2EDuration="35.369294882s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.410610966 +0000 UTC m=+865.761250928" lastFinishedPulling="2025-10-01 16:01:40.18446551 +0000 UTC m=+889.535105512" observedRunningTime="2025-10-01 16:01:49.117439189 +0000 UTC m=+898.468079151" watchObservedRunningTime="2025-10-01 16:01:49.369294882 +0000 UTC m=+898.719934844" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.395374 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" podStartSLOduration=3.304393571 podStartE2EDuration="35.395346716s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.174398389 +0000 UTC m=+865.525038351" lastFinishedPulling="2025-10-01 16:01:48.265351534 +0000 UTC m=+897.615991496" observedRunningTime="2025-10-01 16:01:49.395294644 +0000 UTC m=+898.745934606" watchObservedRunningTime="2025-10-01 16:01:49.395346716 +0000 UTC m=+898.745986678" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.434418 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" podStartSLOduration=12.851807118 podStartE2EDuration="35.434396021s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:17.601798565 +0000 UTC m=+866.952438527" lastFinishedPulling="2025-10-01 16:01:40.184387438 +0000 UTC m=+889.535027430" observedRunningTime="2025-10-01 16:01:49.431175999 +0000 UTC m=+898.781815961" watchObservedRunningTime="2025-10-01 16:01:49.434396021 +0000 UTC m=+898.785035983" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.463801 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" podStartSLOduration=3.5823404290000003 podStartE2EDuration="35.4637815s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.3943078 +0000 UTC m=+865.744947762" lastFinishedPulling="2025-10-01 16:01:48.275748871 +0000 UTC m=+897.626388833" observedRunningTime="2025-10-01 16:01:49.459333723 +0000 UTC m=+898.809973685" watchObservedRunningTime="2025-10-01 16:01:49.4637815 +0000 UTC m=+898.814421462" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.485797 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" podStartSLOduration=3.817460844 podStartE2EDuration="35.485780869s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.233441446 +0000 UTC m=+865.584081408" lastFinishedPulling="2025-10-01 16:01:47.901761471 +0000 UTC m=+897.252401433" observedRunningTime="2025-10-01 16:01:49.483401571 +0000 UTC m=+898.834041523" watchObservedRunningTime="2025-10-01 16:01:49.485780869 +0000 UTC m=+898.836420831" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.504415 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" podStartSLOduration=3.542991426 podStartE2EDuration="35.50439643s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.64079877 +0000 UTC m=+865.991438732" lastFinishedPulling="2025-10-01 16:01:48.602203774 +0000 UTC m=+897.952843736" observedRunningTime="2025-10-01 16:01:49.50437458 +0000 UTC m=+898.855014542" watchObservedRunningTime="2025-10-01 16:01:49.50439643 +0000 UTC m=+898.855036392" Oct 01 16:01:49 crc kubenswrapper[4688]: I1001 16:01:49.522907 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" podStartSLOduration=2.539721993 podStartE2EDuration="35.522889649s" podCreationTimestamp="2025-10-01 16:01:14 +0000 UTC" firstStartedPulling="2025-10-01 16:01:16.007652067 +0000 UTC m=+865.358292029" lastFinishedPulling="2025-10-01 16:01:48.990819723 +0000 UTC m=+898.341459685" observedRunningTime="2025-10-01 16:01:49.52225073 +0000 UTC m=+898.872890692" watchObservedRunningTime="2025-10-01 16:01:49.522889649 +0000 UTC m=+898.873529611" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.616549 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-65bw4" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.721726 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-8pnfq" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.741473 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-548l2" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.763569 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-bh7tx" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.790829 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5qm6b" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.907497 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-85595f4675-l62s8" Oct 01 16:01:54 crc kubenswrapper[4688]: I1001 16:01:54.944654 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-x8lrd" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.000407 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-gbg2c" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.044234 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-mbsl5" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.054315 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.054377 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.107930 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-h7xkj" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.118438 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-9ppnb" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.170919 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-b8dfc" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.193445 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-rvtpg" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.256063 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-g6725" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.322813 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-44nm7" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.352144 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-7blrb" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.372897 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-tp7l5" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.467914 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-6ccdl" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.581218 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-f4v99" Oct 01 16:01:55 crc kubenswrapper[4688]: I1001 16:01:55.670920 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-27q6x" Oct 01 16:01:56 crc kubenswrapper[4688]: I1001 16:01:56.833357 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-77b9676b8cs62rh" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.369679 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.371752 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.373904 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.379255 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.379312 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-x7kj8" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.379574 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.386638 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.386699 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsgzc\" (UniqueName: \"kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.388016 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.451682 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.452888 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.455164 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.465119 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.493179 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.493231 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsgzc\" (UniqueName: \"kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.494336 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.513911 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsgzc\" (UniqueName: \"kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc\") pod \"dnsmasq-dns-675f4bcbfc-sqsgb\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.599410 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stx2t\" (UniqueName: \"kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.599533 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.599555 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.695858 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.700533 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.700578 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.700634 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stx2t\" (UniqueName: \"kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.701358 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.701379 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.724177 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stx2t\" (UniqueName: \"kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t\") pod \"dnsmasq-dns-78dd6ddcc-s2jgt\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:12 crc kubenswrapper[4688]: I1001 16:02:12.769950 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:13 crc kubenswrapper[4688]: I1001 16:02:13.243286 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:13 crc kubenswrapper[4688]: I1001 16:02:13.262558 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:02:13 crc kubenswrapper[4688]: I1001 16:02:13.323071 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:13 crc kubenswrapper[4688]: I1001 16:02:13.586276 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" event={"ID":"849b7166-55d1-4760-915e-8f6336fe0178","Type":"ContainerStarted","Data":"0812b82c72bc1f31f588180cf054ebf54949625ba936d4825663590775c3cefc"} Oct 01 16:02:13 crc kubenswrapper[4688]: I1001 16:02:13.587539 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" event={"ID":"3c79a466-2bef-4ff6-be31-a524290b3cdb","Type":"ContainerStarted","Data":"cc5b53e0b17ad011c0ca5eff6568cbafd17974728030399f46e5edf327550c0f"} Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.565326 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.599365 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.600886 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.616678 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.647016 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9vwb\" (UniqueName: \"kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.647065 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.647108 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.749475 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9vwb\" (UniqueName: \"kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.749555 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.749607 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.750542 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.751726 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.782605 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9vwb\" (UniqueName: \"kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb\") pod \"dnsmasq-dns-666b6646f7-vbxv2\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.922014 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.945138 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.968360 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.969868 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:15 crc kubenswrapper[4688]: I1001 16:02:15.993589 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.061159 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7npsf\" (UniqueName: \"kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.061858 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.061909 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.166025 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.166116 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7npsf\" (UniqueName: \"kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.166163 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.166988 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.167050 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.210026 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7npsf\" (UniqueName: \"kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf\") pod \"dnsmasq-dns-57d769cc4f-mknxc\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.390119 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.716038 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.763351 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.764601 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.771384 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mdzlh" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.771566 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.774737 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.777963 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.777988 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.779315 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.779436 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.782232 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881297 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881357 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881400 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881432 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqwxb\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881461 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881490 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881507 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881535 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881562 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881578 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.881597 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982758 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982818 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982841 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982859 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982883 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982900 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982922 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982950 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.982980 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.983006 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.983021 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqwxb\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.983276 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.984414 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.984543 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.985056 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.985724 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.985895 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.989497 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.989533 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.990476 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:16 crc kubenswrapper[4688]: I1001 16:02:16.997914 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqwxb\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.004099 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.007898 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " pod="openstack/rabbitmq-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.042292 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.139941 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.155506 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.157266 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.159464 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.159809 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tmrnr" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.160078 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.160441 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.160783 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.160974 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.162730 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.185841 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291199 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndqkd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291255 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291286 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291312 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291343 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291381 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291398 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291420 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291442 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291457 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.291480 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392336 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndqkd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392393 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392423 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392445 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392479 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392531 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392555 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392590 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392621 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392643 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.392673 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.393861 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.394576 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.395326 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.399213 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.399451 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.399917 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.400593 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.400794 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.400805 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.405500 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.414056 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndqkd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.449668 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.515734 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.638394 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" event={"ID":"e6df2e13-2cbf-4c51-96e2-b10563a6e189","Type":"ContainerStarted","Data":"c089e4093782d205ae53119279a79fce79e87250fa64fecbf5f65d2a866f22bb"} Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.639515 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" event={"ID":"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e","Type":"ContainerStarted","Data":"285f0b504b191a373e34f093ba67ecf0375b50a3b92c9d17ac2e91f6df6f06d1"} Oct 01 16:02:17 crc kubenswrapper[4688]: I1001 16:02:17.663056 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:02:17 crc kubenswrapper[4688]: W1001 16:02:17.680472 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0372e042_2565_4498_8884_452f6e0bd0f1.slice/crio-0c08901b48369fcd0d0be8e38b86f17709b966518249b6a7214d13123faab7e2 WatchSource:0}: Error finding container 0c08901b48369fcd0d0be8e38b86f17709b966518249b6a7214d13123faab7e2: Status 404 returned error can't find the container with id 0c08901b48369fcd0d0be8e38b86f17709b966518249b6a7214d13123faab7e2 Oct 01 16:02:18 crc kubenswrapper[4688]: I1001 16:02:18.003003 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:02:18 crc kubenswrapper[4688]: W1001 16:02:18.041510 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51ba652f_dca6_4c1b_9d78_1ac10f1bcb3c.slice/crio-146617fef1bc46b28b0ee849ca910185153616c26c899173704406e6f5fb6f24 WatchSource:0}: Error finding container 146617fef1bc46b28b0ee849ca910185153616c26c899173704406e6f5fb6f24: Status 404 returned error can't find the container with id 146617fef1bc46b28b0ee849ca910185153616c26c899173704406e6f5fb6f24 Oct 01 16:02:18 crc kubenswrapper[4688]: I1001 16:02:18.691347 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerStarted","Data":"146617fef1bc46b28b0ee849ca910185153616c26c899173704406e6f5fb6f24"} Oct 01 16:02:18 crc kubenswrapper[4688]: I1001 16:02:18.715310 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerStarted","Data":"0c08901b48369fcd0d0be8e38b86f17709b966518249b6a7214d13123faab7e2"} Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.652092 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.654283 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.660333 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.726034 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-z4lcm" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.726330 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.730902 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.731142 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.731320 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.735865 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843081 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843149 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-kolla-config\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843223 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-secrets\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843256 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843357 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843455 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxnvk\" (UniqueName: \"kubernetes.io/projected/65ea9633-b867-4c49-8fb3-471793344f3b-kube-api-access-gxnvk\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843481 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-default\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843502 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.843576 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.880892 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.882624 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.887781 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-sb5h7" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.888052 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.888278 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.888394 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.902725 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944383 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-kolla-config\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944432 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-secrets\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944452 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944478 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944514 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxnvk\" (UniqueName: \"kubernetes.io/projected/65ea9633-b867-4c49-8fb3-471793344f3b-kube-api-access-gxnvk\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944554 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-default\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944569 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944596 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.944631 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.945490 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-kolla-config\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.947238 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.947381 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.948361 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.948614 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/65ea9633-b867-4c49-8fb3-471793344f3b-config-data-default\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.963118 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-secrets\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.963255 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.963998 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ea9633-b867-4c49-8fb3-471793344f3b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.966041 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxnvk\" (UniqueName: \"kubernetes.io/projected/65ea9633-b867-4c49-8fb3-471793344f3b-kube-api-access-gxnvk\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:19 crc kubenswrapper[4688]: I1001 16:02:19.966318 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"65ea9633-b867-4c49-8fb3-471793344f3b\") " pod="openstack/openstack-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046405 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t28z9\" (UniqueName: \"kubernetes.io/projected/81865451-e9a4-4f89-aa96-eb345800ffe3-kube-api-access-t28z9\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046459 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046495 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046572 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046624 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046644 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046672 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046698 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.046719 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.058377 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147711 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147755 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t28z9\" (UniqueName: \"kubernetes.io/projected/81865451-e9a4-4f89-aa96-eb345800ffe3-kube-api-access-t28z9\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147783 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147814 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147862 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147911 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147935 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147953 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.147973 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.148925 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.148932 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.149152 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.149325 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.149610 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/81865451-e9a4-4f89-aa96-eb345800ffe3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.153701 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.155786 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.160076 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81865451-e9a4-4f89-aa96-eb345800ffe3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.194709 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t28z9\" (UniqueName: \"kubernetes.io/projected/81865451-e9a4-4f89-aa96-eb345800ffe3-kube-api-access-t28z9\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.197693 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"81865451-e9a4-4f89-aa96-eb345800ffe3\") " pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.211009 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.285709 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.286707 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.290501 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.294256 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.294750 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jvs94" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.308124 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.452175 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.457744 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.457795 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-config-data\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.457866 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjxzv\" (UniqueName: \"kubernetes.io/projected/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kube-api-access-kjxzv\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.457944 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.457973 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kolla-config\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.559824 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.559864 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-config-data\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.559894 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjxzv\" (UniqueName: \"kubernetes.io/projected/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kube-api-access-kjxzv\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.559939 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.559960 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kolla-config\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.560671 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kolla-config\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.560936 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-config-data\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.567564 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.568107 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.578171 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjxzv\" (UniqueName: \"kubernetes.io/projected/3720d335-ad5b-41cf-9acb-7f9b49fb2c6c-kube-api-access-kjxzv\") pod \"memcached-0\" (UID: \"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c\") " pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.628302 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.759039 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"65ea9633-b867-4c49-8fb3-471793344f3b","Type":"ContainerStarted","Data":"3d0745be94a0b1e653a6da2b8d2ef6ce3b6a55c6347b767ffd62aed6e6120cdc"} Oct 01 16:02:20 crc kubenswrapper[4688]: I1001 16:02:20.850757 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 16:02:20 crc kubenswrapper[4688]: W1001 16:02:20.853960 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81865451_e9a4_4f89_aa96_eb345800ffe3.slice/crio-4f13c95406965f5b3ece28056ac8b92c48730e7d76dba88d7c06368f5e05b4da WatchSource:0}: Error finding container 4f13c95406965f5b3ece28056ac8b92c48730e7d76dba88d7c06368f5e05b4da: Status 404 returned error can't find the container with id 4f13c95406965f5b3ece28056ac8b92c48730e7d76dba88d7c06368f5e05b4da Oct 01 16:02:21 crc kubenswrapper[4688]: I1001 16:02:21.060113 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 16:02:21 crc kubenswrapper[4688]: W1001 16:02:21.061928 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3720d335_ad5b_41cf_9acb_7f9b49fb2c6c.slice/crio-2ff97c6bcbcb739ae88617a46c75f1b4226d6120e919f3bd453ca82489ac72bd WatchSource:0}: Error finding container 2ff97c6bcbcb739ae88617a46c75f1b4226d6120e919f3bd453ca82489ac72bd: Status 404 returned error can't find the container with id 2ff97c6bcbcb739ae88617a46c75f1b4226d6120e919f3bd453ca82489ac72bd Oct 01 16:02:21 crc kubenswrapper[4688]: I1001 16:02:21.764876 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c","Type":"ContainerStarted","Data":"2ff97c6bcbcb739ae88617a46c75f1b4226d6120e919f3bd453ca82489ac72bd"} Oct 01 16:02:21 crc kubenswrapper[4688]: I1001 16:02:21.765622 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81865451-e9a4-4f89-aa96-eb345800ffe3","Type":"ContainerStarted","Data":"4f13c95406965f5b3ece28056ac8b92c48730e7d76dba88d7c06368f5e05b4da"} Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.147849 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.148776 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.152193 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-j9bxs" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.178467 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.296563 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjx74\" (UniqueName: \"kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74\") pod \"kube-state-metrics-0\" (UID: \"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5\") " pod="openstack/kube-state-metrics-0" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.398053 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjx74\" (UniqueName: \"kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74\") pod \"kube-state-metrics-0\" (UID: \"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5\") " pod="openstack/kube-state-metrics-0" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.414136 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjx74\" (UniqueName: \"kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74\") pod \"kube-state-metrics-0\" (UID: \"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5\") " pod="openstack/kube-state-metrics-0" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.465430 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:02:22 crc kubenswrapper[4688]: I1001 16:02:22.939733 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:02:22 crc kubenswrapper[4688]: W1001 16:02:22.946148 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab594f4_2b07_4d58_bb14_b2ef7dbd59c5.slice/crio-97118b400b502785273900cb1191594ce1da9a9d1869d8b80705aecbcc4a12f1 WatchSource:0}: Error finding container 97118b400b502785273900cb1191594ce1da9a9d1869d8b80705aecbcc4a12f1: Status 404 returned error can't find the container with id 97118b400b502785273900cb1191594ce1da9a9d1869d8b80705aecbcc4a12f1 Oct 01 16:02:23 crc kubenswrapper[4688]: I1001 16:02:23.788206 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5","Type":"ContainerStarted","Data":"97118b400b502785273900cb1191594ce1da9a9d1869d8b80705aecbcc4a12f1"} Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.052830 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.052889 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.052937 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.053649 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.053703 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd" gracePeriod=600 Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.823572 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd" exitCode=0 Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.823992 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd"} Oct 01 16:02:25 crc kubenswrapper[4688]: I1001 16:02:25.824088 4688 scope.go:117] "RemoveContainer" containerID="3d2b36139192ee9b7d99f598809418a816b48a40755bac8424ac57e70df72c94" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.111763 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.114678 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.122882 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.124176 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f726v" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.124376 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.124631 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.124987 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.130955 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.265865 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.265934 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266108 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266208 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266232 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266256 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266276 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-config\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.266292 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7k2\" (UniqueName: \"kubernetes.io/projected/79a0889a-6692-43c9-928b-13e06ccd6ca1-kube-api-access-pd7k2\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367837 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367884 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367913 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367933 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-config\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367949 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7k2\" (UniqueName: \"kubernetes.io/projected/79a0889a-6692-43c9-928b-13e06ccd6ca1-kube-api-access-pd7k2\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.367979 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.368017 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.368069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.369375 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.370184 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79a0889a-6692-43c9-928b-13e06ccd6ca1-config\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.370386 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.371970 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.378124 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.387313 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.403684 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79a0889a-6692-43c9-928b-13e06ccd6ca1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.409065 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.416988 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7k2\" (UniqueName: \"kubernetes.io/projected/79a0889a-6692-43c9-928b-13e06ccd6ca1-kube-api-access-pd7k2\") pod \"ovsdbserver-nb-0\" (UID: \"79a0889a-6692-43c9-928b-13e06ccd6ca1\") " pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.448543 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.740325 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-9wpf7"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.741857 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.743806 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cb9nz" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.744322 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.749789 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ts722"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.751087 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.753207 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.766273 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ts722"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.771763 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9wpf7"] Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.881553 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.881612 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-lib\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.881865 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ab65b16-215b-46d3-9fd7-98982427fffa-scripts\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.882030 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-log-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.882391 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-etc-ovs\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.882620 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.882823 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-combined-ca-bundle\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883265 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-run\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883295 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4559e8a-674a-427d-a9fb-145f94ff3b58-scripts\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883381 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmcxw\" (UniqueName: \"kubernetes.io/projected/3ab65b16-215b-46d3-9fd7-98982427fffa-kube-api-access-wmcxw\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883663 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z9bl\" (UniqueName: \"kubernetes.io/projected/e4559e8a-674a-427d-a9fb-145f94ff3b58-kube-api-access-5z9bl\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883845 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-ovn-controller-tls-certs\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.883872 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-log\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.987718 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.988063 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-lib\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.988218 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ab65b16-215b-46d3-9fd7-98982427fffa-scripts\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.988363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-log-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.988926 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-etc-ovs\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.988974 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989003 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-combined-ca-bundle\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989071 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-run\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989088 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4559e8a-674a-427d-a9fb-145f94ff3b58-scripts\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989110 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmcxw\" (UniqueName: \"kubernetes.io/projected/3ab65b16-215b-46d3-9fd7-98982427fffa-kube-api-access-wmcxw\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989150 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z9bl\" (UniqueName: \"kubernetes.io/projected/e4559e8a-674a-427d-a9fb-145f94ff3b58-kube-api-access-5z9bl\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989178 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-ovn-controller-tls-certs\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.989197 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-log\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.990677 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ab65b16-215b-46d3-9fd7-98982427fffa-scripts\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.990981 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-lib\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.991025 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-etc-ovs\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.991164 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.991263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-log-ovn\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.991368 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-log\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.991944 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3ab65b16-215b-46d3-9fd7-98982427fffa-var-run\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:26 crc kubenswrapper[4688]: I1001 16:02:26.992053 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4559e8a-674a-427d-a9fb-145f94ff3b58-var-run\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:26.997609 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-ovn-controller-tls-certs\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.001402 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4559e8a-674a-427d-a9fb-145f94ff3b58-combined-ca-bundle\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.002518 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4559e8a-674a-427d-a9fb-145f94ff3b58-scripts\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.005924 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmcxw\" (UniqueName: \"kubernetes.io/projected/3ab65b16-215b-46d3-9fd7-98982427fffa-kube-api-access-wmcxw\") pod \"ovn-controller-ovs-9wpf7\" (UID: \"3ab65b16-215b-46d3-9fd7-98982427fffa\") " pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.006955 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z9bl\" (UniqueName: \"kubernetes.io/projected/e4559e8a-674a-427d-a9fb-145f94ff3b58-kube-api-access-5z9bl\") pod \"ovn-controller-ts722\" (UID: \"e4559e8a-674a-427d-a9fb-145f94ff3b58\") " pod="openstack/ovn-controller-ts722" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.093848 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:27 crc kubenswrapper[4688]: I1001 16:02:27.125606 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.643980 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.646732 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.648493 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.649928 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.650035 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bdbf9" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.651167 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.652249 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735247 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-config\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735291 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735331 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735399 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgbkr\" (UniqueName: \"kubernetes.io/projected/466a92b3-31f5-41fb-8810-76e95dd7a332-kube-api-access-jgbkr\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735437 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735462 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735492 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.735588 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837064 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837133 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837177 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837205 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-config\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837235 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837260 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837316 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgbkr\" (UniqueName: \"kubernetes.io/projected/466a92b3-31f5-41fb-8810-76e95dd7a332-kube-api-access-jgbkr\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.837358 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.838673 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.839023 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.839071 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-config\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.840778 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/466a92b3-31f5-41fb-8810-76e95dd7a332-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.844799 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.846925 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.863236 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgbkr\" (UniqueName: \"kubernetes.io/projected/466a92b3-31f5-41fb-8810-76e95dd7a332-kube-api-access-jgbkr\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.868231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/466a92b3-31f5-41fb-8810-76e95dd7a332-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.878733 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"466a92b3-31f5-41fb-8810-76e95dd7a332\") " pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:29 crc kubenswrapper[4688]: I1001 16:02:29.966351 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.063809 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.065786 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqwxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0372e042-2565-4498-8884-452f6e0bd0f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.067001 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.092192 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.092376 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ndqkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.093550 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.812101 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.812479 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w9vwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-vbxv2_openstack(e6df2e13-2cbf-4c51-96e2-b10563a6e189): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.813785 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" podUID="e6df2e13-2cbf-4c51-96e2-b10563a6e189" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.849653 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.849821 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jsgzc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-sqsgb_openstack(849b7166-55d1-4760-915e-8f6336fe0178): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.850994 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" podUID="849b7166-55d1-4760-915e-8f6336fe0178" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.948857 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.953507 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" Oct 01 16:02:42 crc kubenswrapper[4688]: E1001 16:02:42.954462 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" podUID="e6df2e13-2cbf-4c51-96e2-b10563a6e189" Oct 01 16:02:44 crc kubenswrapper[4688]: E1001 16:02:44.674544 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 01 16:02:44 crc kubenswrapper[4688]: E1001 16:02:44.675027 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t28z9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(81865451-e9a4-4f89-aa96-eb345800ffe3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:44 crc kubenswrapper[4688]: E1001 16:02:44.676149 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="81865451-e9a4-4f89-aa96-eb345800ffe3" Oct 01 16:02:44 crc kubenswrapper[4688]: E1001 16:02:44.962623 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="81865451-e9a4-4f89-aa96-eb345800ffe3" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.351558 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.351731 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n64bh689h594h5d4h556h5ddh77h59bh574h66fh66ch59h7hc5h5d7hbfh654h557h565h647hc5h56bh6h5d4h698h58h568h65h664hd4h67h57dq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjxzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(3720d335-ad5b-41cf-9acb-7f9b49fb2c6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.352918 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="3720d335-ad5b-41cf-9acb-7f9b49fb2c6c" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.358247 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.358294 4688 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.358414 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zjx74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" logger="UnhandledError" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.359582 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.371366 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.371510 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7npsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-mknxc_openstack(f855fd46-0733-4d7f-b17f-eba1ed6e2b0e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.373261 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" podUID="f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.503172 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.503645 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stx2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-s2jgt_openstack(3c79a466-2bef-4ff6-be31-a524290b3cdb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.504731 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" podUID="3c79a466-2bef-4ff6-be31-a524290b3cdb" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.552732 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.632439 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsgzc\" (UniqueName: \"kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc\") pod \"849b7166-55d1-4760-915e-8f6336fe0178\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.632487 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config\") pod \"849b7166-55d1-4760-915e-8f6336fe0178\" (UID: \"849b7166-55d1-4760-915e-8f6336fe0178\") " Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.633370 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config" (OuterVolumeSpecName: "config") pod "849b7166-55d1-4760-915e-8f6336fe0178" (UID: "849b7166-55d1-4760-915e-8f6336fe0178"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.640292 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc" (OuterVolumeSpecName: "kube-api-access-jsgzc") pod "849b7166-55d1-4760-915e-8f6336fe0178" (UID: "849b7166-55d1-4760-915e-8f6336fe0178"). InnerVolumeSpecName "kube-api-access-jsgzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.734689 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsgzc\" (UniqueName: \"kubernetes.io/projected/849b7166-55d1-4760-915e-8f6336fe0178-kube-api-access-jsgzc\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.734723 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/849b7166-55d1-4760-915e-8f6336fe0178-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.918998 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ts722"] Oct 01 16:02:45 crc kubenswrapper[4688]: W1001 16:02:45.927322 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4559e8a_674a_427d_a9fb_145f94ff3b58.slice/crio-b4f26a51207da6379298b72bbe674b3888928f81971a8e0bfa4c95219388f2dd WatchSource:0}: Error finding container b4f26a51207da6379298b72bbe674b3888928f81971a8e0bfa4c95219388f2dd: Status 404 returned error can't find the container with id b4f26a51207da6379298b72bbe674b3888928f81971a8e0bfa4c95219388f2dd Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.967552 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"65ea9633-b867-4c49-8fb3-471793344f3b","Type":"ContainerStarted","Data":"02e7d0d856a6f889d1205506b8ffbd9f9e79900e5343000d2de1cc7663300ef3"} Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.969171 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722" event={"ID":"e4559e8a-674a-427d-a9fb-145f94ff3b58","Type":"ContainerStarted","Data":"b4f26a51207da6379298b72bbe674b3888928f81971a8e0bfa4c95219388f2dd"} Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.973229 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18"} Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.975449 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" Oct 01 16:02:45 crc kubenswrapper[4688]: I1001 16:02:45.975649 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-sqsgb" event={"ID":"849b7166-55d1-4760-915e-8f6336fe0178","Type":"ContainerDied","Data":"0812b82c72bc1f31f588180cf054ebf54949625ba936d4825663590775c3cefc"} Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.976913 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.977771 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" podUID="f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" Oct 01 16:02:45 crc kubenswrapper[4688]: E1001 16:02:45.977827 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="3720d335-ad5b-41cf-9acb-7f9b49fb2c6c" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.162319 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.173305 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sqsgb"] Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.349837 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.456976 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9wpf7"] Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.457401 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config\") pod \"3c79a466-2bef-4ff6-be31-a524290b3cdb\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.457512 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc\") pod \"3c79a466-2bef-4ff6-be31-a524290b3cdb\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.457559 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stx2t\" (UniqueName: \"kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t\") pod \"3c79a466-2bef-4ff6-be31-a524290b3cdb\" (UID: \"3c79a466-2bef-4ff6-be31-a524290b3cdb\") " Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.458190 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config" (OuterVolumeSpecName: "config") pod "3c79a466-2bef-4ff6-be31-a524290b3cdb" (UID: "3c79a466-2bef-4ff6-be31-a524290b3cdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.458678 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c79a466-2bef-4ff6-be31-a524290b3cdb" (UID: "3c79a466-2bef-4ff6-be31-a524290b3cdb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.464790 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t" (OuterVolumeSpecName: "kube-api-access-stx2t") pod "3c79a466-2bef-4ff6-be31-a524290b3cdb" (UID: "3c79a466-2bef-4ff6-be31-a524290b3cdb"). InnerVolumeSpecName "kube-api-access-stx2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.559546 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.559577 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c79a466-2bef-4ff6-be31-a524290b3cdb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.559587 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stx2t\" (UniqueName: \"kubernetes.io/projected/3c79a466-2bef-4ff6-be31-a524290b3cdb-kube-api-access-stx2t\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.984677 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" event={"ID":"3c79a466-2bef-4ff6-be31-a524290b3cdb","Type":"ContainerDied","Data":"cc5b53e0b17ad011c0ca5eff6568cbafd17974728030399f46e5edf327550c0f"} Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.984697 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s2jgt" Oct 01 16:02:46 crc kubenswrapper[4688]: I1001 16:02:46.986867 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9wpf7" event={"ID":"3ab65b16-215b-46d3-9fd7-98982427fffa","Type":"ContainerStarted","Data":"dd710c04cbef85a8ca5ce58891da1283d2980487bb262f492fd1a0007864de38"} Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.057405 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.062258 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s2jgt"] Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.340771 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.389744 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c79a466-2bef-4ff6-be31-a524290b3cdb" path="/var/lib/kubelet/pods/3c79a466-2bef-4ff6-be31-a524290b3cdb/volumes" Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.390157 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="849b7166-55d1-4760-915e-8f6336fe0178" path="/var/lib/kubelet/pods/849b7166-55d1-4760-915e-8f6336fe0178/volumes" Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.493617 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 16:02:47 crc kubenswrapper[4688]: I1001 16:02:47.997721 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"466a92b3-31f5-41fb-8810-76e95dd7a332","Type":"ContainerStarted","Data":"6a4931e7c10c618bd3f18569c322f3684101711ba5e0f7b72c35a5efc6280c19"} Oct 01 16:02:48 crc kubenswrapper[4688]: W1001 16:02:48.798834 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79a0889a_6692_43c9_928b_13e06ccd6ca1.slice/crio-002752fe144fa41390d534fd0bfaa948096f8d8e714d19ee2bb220af052834ca WatchSource:0}: Error finding container 002752fe144fa41390d534fd0bfaa948096f8d8e714d19ee2bb220af052834ca: Status 404 returned error can't find the container with id 002752fe144fa41390d534fd0bfaa948096f8d8e714d19ee2bb220af052834ca Oct 01 16:02:49 crc kubenswrapper[4688]: I1001 16:02:49.004978 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79a0889a-6692-43c9-928b-13e06ccd6ca1","Type":"ContainerStarted","Data":"002752fe144fa41390d534fd0bfaa948096f8d8e714d19ee2bb220af052834ca"} Oct 01 16:02:49 crc kubenswrapper[4688]: I1001 16:02:49.973878 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-sfjc5"] Oct 01 16:02:49 crc kubenswrapper[4688]: I1001 16:02:49.975653 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:49 crc kubenswrapper[4688]: I1001 16:02:49.982674 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.007582 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sfjc5"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.017199 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-combined-ca-bundle\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.017465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.017719 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3cd14b-091c-461a-8ce5-924f9619a596-config\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.018744 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovn-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.018947 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovs-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.019097 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j9p\" (UniqueName: \"kubernetes.io/projected/9b3cd14b-091c-461a-8ce5-924f9619a596-kube-api-access-f9j9p\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.025290 4688 generic.go:334] "Generic (PLEG): container finished" podID="65ea9633-b867-4c49-8fb3-471793344f3b" containerID="02e7d0d856a6f889d1205506b8ffbd9f9e79900e5343000d2de1cc7663300ef3" exitCode=0 Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.025330 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"65ea9633-b867-4c49-8fb3-471793344f3b","Type":"ContainerDied","Data":"02e7d0d856a6f889d1205506b8ffbd9f9e79900e5343000d2de1cc7663300ef3"} Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123417 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovs-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123505 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j9p\" (UniqueName: \"kubernetes.io/projected/9b3cd14b-091c-461a-8ce5-924f9619a596-kube-api-access-f9j9p\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123649 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-combined-ca-bundle\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123691 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123721 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3cd14b-091c-461a-8ce5-924f9619a596-config\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.123805 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovn-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.124102 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovn-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.124168 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9b3cd14b-091c-461a-8ce5-924f9619a596-ovs-rundir\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.131117 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3cd14b-091c-461a-8ce5-924f9619a596-config\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.134299 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-combined-ca-bundle\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.143239 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b3cd14b-091c-461a-8ce5-924f9619a596-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.170560 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j9p\" (UniqueName: \"kubernetes.io/projected/9b3cd14b-091c-461a-8ce5-924f9619a596-kube-api-access-f9j9p\") pod \"ovn-controller-metrics-sfjc5\" (UID: \"9b3cd14b-091c-461a-8ce5-924f9619a596\") " pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.192516 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.254049 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.262431 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.266112 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.296487 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.302499 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sfjc5" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.327308 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tqsr\" (UniqueName: \"kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.327396 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.327450 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.327516 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.415155 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.431134 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.431954 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.432329 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tqsr\" (UniqueName: \"kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.432650 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.433220 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.433270 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.434027 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.471875 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.473092 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.475134 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.496151 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tqsr\" (UniqueName: \"kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr\") pod \"dnsmasq-dns-5bf47b49b7-9khc8\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.513207 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.552675 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.552780 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.552859 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.552902 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6rht\" (UniqueName: \"kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.552978 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.609549 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.654329 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.654378 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6rht\" (UniqueName: \"kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.654421 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.654469 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.654507 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.655831 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.655859 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.656446 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.656566 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.691298 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6rht\" (UniqueName: \"kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht\") pod \"dnsmasq-dns-8554648995-nxw4l\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.729659 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.833798 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.858546 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc\") pod \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.858629 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7npsf\" (UniqueName: \"kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf\") pod \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.858799 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config\") pod \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\" (UID: \"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e\") " Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.860050 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config" (OuterVolumeSpecName: "config") pod "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" (UID: "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.860064 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" (UID: "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.882420 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf" (OuterVolumeSpecName: "kube-api-access-7npsf") pod "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" (UID: "f855fd46-0733-4d7f-b17f-eba1ed6e2b0e"). InnerVolumeSpecName "kube-api-access-7npsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.960588 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7npsf\" (UniqueName: \"kubernetes.io/projected/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-kube-api-access-7npsf\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.960957 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:50 crc kubenswrapper[4688]: I1001 16:02:50.960970 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.014490 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sfjc5"] Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.051827 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9wpf7" event={"ID":"3ab65b16-215b-46d3-9fd7-98982427fffa","Type":"ContainerStarted","Data":"401201f8b84dc6371f1295c023892ec905ddebbeea137d125379967a27492411"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.078178 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"65ea9633-b867-4c49-8fb3-471793344f3b","Type":"ContainerStarted","Data":"7a41bd0076e1efe4465cca321467e0dace8f1c8fff3b255bdb28a94b422071a5"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.083776 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" event={"ID":"f855fd46-0733-4d7f-b17f-eba1ed6e2b0e","Type":"ContainerDied","Data":"285f0b504b191a373e34f093ba67ecf0375b50a3b92c9d17ac2e91f6df6f06d1"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.083873 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-mknxc" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.088211 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722" event={"ID":"e4559e8a-674a-427d-a9fb-145f94ff3b58","Type":"ContainerStarted","Data":"263e795091a403085eb0d52c7a48e05e7f079f4eb142488b6d0129757220ad11"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.088487 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ts722" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.096379 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sfjc5" event={"ID":"9b3cd14b-091c-461a-8ce5-924f9619a596","Type":"ContainerStarted","Data":"9b89080dfdc3d056e0bd103fb6418349ca585c14b828c363c8b82a0259433472"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.099560 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"466a92b3-31f5-41fb-8810-76e95dd7a332","Type":"ContainerStarted","Data":"bf994f5c5441b600b7f95bc6769ceee608d06c88d67db3523c3fd0c3f33bf1e7"} Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.152940 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.169755546 podStartE2EDuration="33.15292262s" podCreationTimestamp="2025-10-01 16:02:18 +0000 UTC" firstStartedPulling="2025-10-01 16:02:20.460793086 +0000 UTC m=+929.811433048" lastFinishedPulling="2025-10-01 16:02:45.44396016 +0000 UTC m=+954.794600122" observedRunningTime="2025-10-01 16:02:51.111423765 +0000 UTC m=+960.462063727" watchObservedRunningTime="2025-10-01 16:02:51.15292262 +0000 UTC m=+960.503562582" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.155351 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ts722" podStartSLOduration=21.371224738 podStartE2EDuration="25.155341299s" podCreationTimestamp="2025-10-01 16:02:26 +0000 UTC" firstStartedPulling="2025-10-01 16:02:45.929793732 +0000 UTC m=+955.280433694" lastFinishedPulling="2025-10-01 16:02:49.713910283 +0000 UTC m=+959.064550255" observedRunningTime="2025-10-01 16:02:51.145515696 +0000 UTC m=+960.496155658" watchObservedRunningTime="2025-10-01 16:02:51.155341299 +0000 UTC m=+960.505981261" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.211386 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.275355 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9vwb\" (UniqueName: \"kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb\") pod \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.277263 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config\") pod \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.277373 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc\") pod \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\" (UID: \"e6df2e13-2cbf-4c51-96e2-b10563a6e189\") " Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.278347 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config" (OuterVolumeSpecName: "config") pod "e6df2e13-2cbf-4c51-96e2-b10563a6e189" (UID: "e6df2e13-2cbf-4c51-96e2-b10563a6e189"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.278894 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6df2e13-2cbf-4c51-96e2-b10563a6e189" (UID: "e6df2e13-2cbf-4c51-96e2-b10563a6e189"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.279335 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.279397 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6df2e13-2cbf-4c51-96e2-b10563a6e189-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.286407 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb" (OuterVolumeSpecName: "kube-api-access-w9vwb") pod "e6df2e13-2cbf-4c51-96e2-b10563a6e189" (UID: "e6df2e13-2cbf-4c51-96e2-b10563a6e189"). InnerVolumeSpecName "kube-api-access-w9vwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.340114 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.358122 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-mknxc"] Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.444086 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9vwb\" (UniqueName: \"kubernetes.io/projected/e6df2e13-2cbf-4c51-96e2-b10563a6e189-kube-api-access-w9vwb\") on node \"crc\" DevicePath \"\"" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.471740 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f855fd46-0733-4d7f-b17f-eba1ed6e2b0e" path="/var/lib/kubelet/pods/f855fd46-0733-4d7f-b17f-eba1ed6e2b0e/volumes" Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.472230 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:02:51 crc kubenswrapper[4688]: I1001 16:02:51.512181 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:02:51 crc kubenswrapper[4688]: W1001 16:02:51.547713 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79f808de_0fe9_4ed9_aa09_826434cd24bb.slice/crio-2896ecd3183ef7f6423ae8b7d4cb73cecd5ecf5419508d70dfc524f825fcf31d WatchSource:0}: Error finding container 2896ecd3183ef7f6423ae8b7d4cb73cecd5ecf5419508d70dfc524f825fcf31d: Status 404 returned error can't find the container with id 2896ecd3183ef7f6423ae8b7d4cb73cecd5ecf5419508d70dfc524f825fcf31d Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.108869 4688 generic.go:334] "Generic (PLEG): container finished" podID="3ab65b16-215b-46d3-9fd7-98982427fffa" containerID="401201f8b84dc6371f1295c023892ec905ddebbeea137d125379967a27492411" exitCode=0 Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.109021 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9wpf7" event={"ID":"3ab65b16-215b-46d3-9fd7-98982427fffa","Type":"ContainerDied","Data":"401201f8b84dc6371f1295c023892ec905ddebbeea137d125379967a27492411"} Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.111855 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nxw4l" event={"ID":"79f808de-0fe9-4ed9-aa09-826434cd24bb","Type":"ContainerStarted","Data":"2896ecd3183ef7f6423ae8b7d4cb73cecd5ecf5419508d70dfc524f825fcf31d"} Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.113174 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.113157 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" event={"ID":"81eed4a0-35bf-4aa0-b1c0-4208d39946d9","Type":"ContainerStarted","Data":"5622678dc36aca6b7ca61a9b8f3afc85a08cf34817ab62f25d3bfca3a5941d76"} Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.113364 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vbxv2" event={"ID":"e6df2e13-2cbf-4c51-96e2-b10563a6e189","Type":"ContainerDied","Data":"c089e4093782d205ae53119279a79fce79e87250fa64fecbf5f65d2a866f22bb"} Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.207341 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:52 crc kubenswrapper[4688]: I1001 16:02:52.212599 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vbxv2"] Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.124282 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9wpf7" event={"ID":"3ab65b16-215b-46d3-9fd7-98982427fffa","Type":"ContainerStarted","Data":"7c34851a50a441b31fbad2aa9eef6383a98419785e03d48184543d1a93c02f52"} Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.126282 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79a0889a-6692-43c9-928b-13e06ccd6ca1","Type":"ContainerStarted","Data":"6b7b87313f46a67b1ac27690c05229a3926219ad257a8342b3c90401a3da1338"} Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.127599 4688 generic.go:334] "Generic (PLEG): container finished" podID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerID="600adccb09146af7cf9c0aca8bb3648a97d7928cf37a920becba6f8c7919d0fd" exitCode=0 Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.127634 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nxw4l" event={"ID":"79f808de-0fe9-4ed9-aa09-826434cd24bb","Type":"ContainerDied","Data":"600adccb09146af7cf9c0aca8bb3648a97d7928cf37a920becba6f8c7919d0fd"} Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.156085 4688 generic.go:334] "Generic (PLEG): container finished" podID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerID="145f592971ebaf0b81ef0aaf402bae3a0a622ec5d66573c71e320bc343aad6af" exitCode=0 Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.156128 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" event={"ID":"81eed4a0-35bf-4aa0-b1c0-4208d39946d9","Type":"ContainerDied","Data":"145f592971ebaf0b81ef0aaf402bae3a0a622ec5d66573c71e320bc343aad6af"} Oct 01 16:02:53 crc kubenswrapper[4688]: I1001 16:02:53.405424 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6df2e13-2cbf-4c51-96e2-b10563a6e189" path="/var/lib/kubelet/pods/e6df2e13-2cbf-4c51-96e2-b10563a6e189/volumes" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.166937 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9wpf7" event={"ID":"3ab65b16-215b-46d3-9fd7-98982427fffa","Type":"ContainerStarted","Data":"edb1ec1c4f800460ed509d3a81b213d95676ed2d3f563cad175bdb1270dc40a9"} Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.167383 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.167417 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.169203 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" event={"ID":"81eed4a0-35bf-4aa0-b1c0-4208d39946d9","Type":"ContainerStarted","Data":"3109035f93f672a681e589bc4c8df88d99644d70436a88c9858004c8107371a9"} Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.169921 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.172440 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nxw4l" event={"ID":"79f808de-0fe9-4ed9-aa09-826434cd24bb","Type":"ContainerStarted","Data":"449ff1359abc109fda1da7fab8490b8aed89bbd63b67d91ae26cd761e242fa1f"} Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.172581 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.206879 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-9wpf7" podStartSLOduration=24.970576902 podStartE2EDuration="28.206857165s" podCreationTimestamp="2025-10-01 16:02:26 +0000 UTC" firstStartedPulling="2025-10-01 16:02:46.475138968 +0000 UTC m=+955.825778940" lastFinishedPulling="2025-10-01 16:02:49.711419241 +0000 UTC m=+959.062059203" observedRunningTime="2025-10-01 16:02:54.20423261 +0000 UTC m=+963.554872582" watchObservedRunningTime="2025-10-01 16:02:54.206857165 +0000 UTC m=+963.557497127" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.229068 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-nxw4l" podStartSLOduration=3.138768164 podStartE2EDuration="4.229039373s" podCreationTimestamp="2025-10-01 16:02:50 +0000 UTC" firstStartedPulling="2025-10-01 16:02:51.550238865 +0000 UTC m=+960.900878827" lastFinishedPulling="2025-10-01 16:02:52.640510074 +0000 UTC m=+961.991150036" observedRunningTime="2025-10-01 16:02:54.222373102 +0000 UTC m=+963.573013064" watchObservedRunningTime="2025-10-01 16:02:54.229039373 +0000 UTC m=+963.579679335" Oct 01 16:02:54 crc kubenswrapper[4688]: I1001 16:02:54.250137 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" podStartSLOduration=3.073905498 podStartE2EDuration="4.25012011s" podCreationTimestamp="2025-10-01 16:02:50 +0000 UTC" firstStartedPulling="2025-10-01 16:02:51.465748313 +0000 UTC m=+960.816388275" lastFinishedPulling="2025-10-01 16:02:52.641962925 +0000 UTC m=+961.992602887" observedRunningTime="2025-10-01 16:02:54.245012113 +0000 UTC m=+963.595652085" watchObservedRunningTime="2025-10-01 16:02:54.25012011 +0000 UTC m=+963.600760072" Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.204065 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79a0889a-6692-43c9-928b-13e06ccd6ca1","Type":"ContainerStarted","Data":"05de168ab0735f710e318547a9151792c1f5ca517ad4af5cfa555c395cc47455"} Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.205366 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sfjc5" event={"ID":"9b3cd14b-091c-461a-8ce5-924f9619a596","Type":"ContainerStarted","Data":"67f1cddaa9c2c0ac10dd0e897e68113fadf8dca115367727e9a8cc081fed300f"} Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.207629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"466a92b3-31f5-41fb-8810-76e95dd7a332","Type":"ContainerStarted","Data":"6552c0f98b1891fbc8532da596e261f1c70d8380408e987ef12ac4948ff85428"} Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.227583 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=24.484642676 podStartE2EDuration="32.227569934s" podCreationTimestamp="2025-10-01 16:02:25 +0000 UTC" firstStartedPulling="2025-10-01 16:02:48.803796 +0000 UTC m=+958.154435962" lastFinishedPulling="2025-10-01 16:02:56.546723258 +0000 UTC m=+965.897363220" observedRunningTime="2025-10-01 16:02:57.226468402 +0000 UTC m=+966.577108364" watchObservedRunningTime="2025-10-01 16:02:57.227569934 +0000 UTC m=+966.578209896" Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.244789 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-sfjc5" podStartSLOduration=2.799278542 podStartE2EDuration="8.244769199s" podCreationTimestamp="2025-10-01 16:02:49 +0000 UTC" firstStartedPulling="2025-10-01 16:02:51.022539977 +0000 UTC m=+960.373179939" lastFinishedPulling="2025-10-01 16:02:56.468030624 +0000 UTC m=+965.818670596" observedRunningTime="2025-10-01 16:02:57.243929175 +0000 UTC m=+966.594569137" watchObservedRunningTime="2025-10-01 16:02:57.244769199 +0000 UTC m=+966.595409181" Oct 01 16:02:57 crc kubenswrapper[4688]: I1001 16:02:57.300266 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.204953985 podStartE2EDuration="29.300248176s" podCreationTimestamp="2025-10-01 16:02:28 +0000 UTC" firstStartedPulling="2025-10-01 16:02:47.347813995 +0000 UTC m=+956.698453957" lastFinishedPulling="2025-10-01 16:02:56.443108186 +0000 UTC m=+965.793748148" observedRunningTime="2025-10-01 16:02:57.299242087 +0000 UTC m=+966.649882049" watchObservedRunningTime="2025-10-01 16:02:57.300248176 +0000 UTC m=+966.650888138" Oct 01 16:02:58 crc kubenswrapper[4688]: I1001 16:02:58.227172 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerStarted","Data":"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d"} Oct 01 16:02:58 crc kubenswrapper[4688]: I1001 16:02:58.231576 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81865451-e9a4-4f89-aa96-eb345800ffe3","Type":"ContainerStarted","Data":"40d79a868229c57fcf20ce7779ea879909827e8fb97f7f60731ac6de3a3cf871"} Oct 01 16:02:58 crc kubenswrapper[4688]: I1001 16:02:58.236051 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerStarted","Data":"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567"} Oct 01 16:02:59 crc kubenswrapper[4688]: I1001 16:02:59.449419 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:59 crc kubenswrapper[4688]: I1001 16:02:59.505833 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 01 16:02:59 crc kubenswrapper[4688]: I1001 16:02:59.968022 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 01 16:02:59 crc kubenswrapper[4688]: I1001 16:02:59.968087 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.058672 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.058977 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.072237 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.193228 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.248960 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.286909 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.292881 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.297143 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.593823 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.602311 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.606997 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.607256 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.607361 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.607461 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6jj8g" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.614677 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.619453 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.662280 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7f469"] Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.663153 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7f469" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.677742 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7f469"] Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.711362 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.711979 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-config\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712157 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712290 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-scripts\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712453 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rwv8\" (UniqueName: \"kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8\") pod \"placement-db-create-7f469\" (UID: \"88217dd8-e3f4-4a77-8960-80980cda5e6a\") " pod="openstack/placement-db-create-7f469" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712589 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712736 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.712871 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6shv5\" (UniqueName: \"kubernetes.io/projected/33e5a114-5ed4-4df4-88ca-908340de2204-kube-api-access-6shv5\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814661 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814712 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-scripts\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814761 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rwv8\" (UniqueName: \"kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8\") pod \"placement-db-create-7f469\" (UID: \"88217dd8-e3f4-4a77-8960-80980cda5e6a\") " pod="openstack/placement-db-create-7f469" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814786 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814826 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814862 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6shv5\" (UniqueName: \"kubernetes.io/projected/33e5a114-5ed4-4df4-88ca-908340de2204-kube-api-access-6shv5\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814891 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.814944 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-config\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.815100 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.815975 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-config\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.825264 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.825829 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.830678 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33e5a114-5ed4-4df4-88ca-908340de2204-scripts\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.834189 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e5a114-5ed4-4df4-88ca-908340de2204-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.835012 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6shv5\" (UniqueName: \"kubernetes.io/projected/33e5a114-5ed4-4df4-88ca-908340de2204-kube-api-access-6shv5\") pod \"ovn-northd-0\" (UID: \"33e5a114-5ed4-4df4-88ca-908340de2204\") " pod="openstack/ovn-northd-0" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.843222 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.848162 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rwv8\" (UniqueName: \"kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8\") pod \"placement-db-create-7f469\" (UID: \"88217dd8-e3f4-4a77-8960-80980cda5e6a\") " pod="openstack/placement-db-create-7f469" Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.910633 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:03:00 crc kubenswrapper[4688]: I1001 16:03:00.923951 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 16:03:01 crc kubenswrapper[4688]: I1001 16:03:01.019817 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7f469" Oct 01 16:03:01 crc kubenswrapper[4688]: I1001 16:03:01.254108 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="dnsmasq-dns" containerID="cri-o://3109035f93f672a681e589bc4c8df88d99644d70436a88c9858004c8107371a9" gracePeriod=10 Oct 01 16:03:01 crc kubenswrapper[4688]: I1001 16:03:01.507753 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 16:03:01 crc kubenswrapper[4688]: W1001 16:03:01.514930 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33e5a114_5ed4_4df4_88ca_908340de2204.slice/crio-d25f65fc53fefd8e48f12ed6084231d3857f0210d8af1297a75551fac93e87fe WatchSource:0}: Error finding container d25f65fc53fefd8e48f12ed6084231d3857f0210d8af1297a75551fac93e87fe: Status 404 returned error can't find the container with id d25f65fc53fefd8e48f12ed6084231d3857f0210d8af1297a75551fac93e87fe Oct 01 16:03:01 crc kubenswrapper[4688]: I1001 16:03:01.594507 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7f469"] Oct 01 16:03:01 crc kubenswrapper[4688]: W1001 16:03:01.595312 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88217dd8_e3f4_4a77_8960_80980cda5e6a.slice/crio-27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6 WatchSource:0}: Error finding container 27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6: Status 404 returned error can't find the container with id 27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6 Oct 01 16:03:02 crc kubenswrapper[4688]: I1001 16:03:02.261467 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"33e5a114-5ed4-4df4-88ca-908340de2204","Type":"ContainerStarted","Data":"d25f65fc53fefd8e48f12ed6084231d3857f0210d8af1297a75551fac93e87fe"} Oct 01 16:03:02 crc kubenswrapper[4688]: I1001 16:03:02.263107 4688 generic.go:334] "Generic (PLEG): container finished" podID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerID="3109035f93f672a681e589bc4c8df88d99644d70436a88c9858004c8107371a9" exitCode=0 Oct 01 16:03:02 crc kubenswrapper[4688]: I1001 16:03:02.263180 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" event={"ID":"81eed4a0-35bf-4aa0-b1c0-4208d39946d9","Type":"ContainerDied","Data":"3109035f93f672a681e589bc4c8df88d99644d70436a88c9858004c8107371a9"} Oct 01 16:03:02 crc kubenswrapper[4688]: I1001 16:03:02.266554 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7f469" event={"ID":"88217dd8-e3f4-4a77-8960-80980cda5e6a","Type":"ContainerStarted","Data":"669b5751d549abf566560b9bb60f58c3f0b197455f9e49bded7b00a72df35f21"} Oct 01 16:03:02 crc kubenswrapper[4688]: I1001 16:03:02.266582 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7f469" event={"ID":"88217dd8-e3f4-4a77-8960-80980cda5e6a","Type":"ContainerStarted","Data":"27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6"} Oct 01 16:03:03 crc kubenswrapper[4688]: I1001 16:03:03.297199 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-7f469" podStartSLOduration=3.297175445 podStartE2EDuration="3.297175445s" podCreationTimestamp="2025-10-01 16:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:03.289848074 +0000 UTC m=+972.640488036" watchObservedRunningTime="2025-10-01 16:03:03.297175445 +0000 UTC m=+972.647815447" Oct 01 16:03:04 crc kubenswrapper[4688]: I1001 16:03:04.913490 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.086019 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tqsr\" (UniqueName: \"kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr\") pod \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.086134 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb\") pod \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.086183 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config\") pod \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.086228 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc\") pod \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\" (UID: \"81eed4a0-35bf-4aa0-b1c0-4208d39946d9\") " Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.094973 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr" (OuterVolumeSpecName: "kube-api-access-4tqsr") pod "81eed4a0-35bf-4aa0-b1c0-4208d39946d9" (UID: "81eed4a0-35bf-4aa0-b1c0-4208d39946d9"). InnerVolumeSpecName "kube-api-access-4tqsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.126849 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81eed4a0-35bf-4aa0-b1c0-4208d39946d9" (UID: "81eed4a0-35bf-4aa0-b1c0-4208d39946d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.128968 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config" (OuterVolumeSpecName: "config") pod "81eed4a0-35bf-4aa0-b1c0-4208d39946d9" (UID: "81eed4a0-35bf-4aa0-b1c0-4208d39946d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.130990 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81eed4a0-35bf-4aa0-b1c0-4208d39946d9" (UID: "81eed4a0-35bf-4aa0-b1c0-4208d39946d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.188051 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tqsr\" (UniqueName: \"kubernetes.io/projected/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-kube-api-access-4tqsr\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.188074 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.188085 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.188093 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81eed4a0-35bf-4aa0-b1c0-4208d39946d9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.292761 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" event={"ID":"81eed4a0-35bf-4aa0-b1c0-4208d39946d9","Type":"ContainerDied","Data":"5622678dc36aca6b7ca61a9b8f3afc85a08cf34817ab62f25d3bfca3a5941d76"} Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.292839 4688 scope.go:117] "RemoveContainer" containerID="3109035f93f672a681e589bc4c8df88d99644d70436a88c9858004c8107371a9" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.292842 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9khc8" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.296147 4688 generic.go:334] "Generic (PLEG): container finished" podID="88217dd8-e3f4-4a77-8960-80980cda5e6a" containerID="669b5751d549abf566560b9bb60f58c3f0b197455f9e49bded7b00a72df35f21" exitCode=0 Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.296202 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7f469" event={"ID":"88217dd8-e3f4-4a77-8960-80980cda5e6a","Type":"ContainerDied","Data":"669b5751d549abf566560b9bb60f58c3f0b197455f9e49bded7b00a72df35f21"} Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.321203 4688 scope.go:117] "RemoveContainer" containerID="145f592971ebaf0b81ef0aaf402bae3a0a622ec5d66573c71e320bc343aad6af" Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.354428 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.364223 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9khc8"] Oct 01 16:03:05 crc kubenswrapper[4688]: I1001 16:03:05.392715 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" path="/var/lib/kubelet/pods/81eed4a0-35bf-4aa0-b1c0-4208d39946d9/volumes" Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.309128 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3720d335-ad5b-41cf-9acb-7f9b49fb2c6c","Type":"ContainerStarted","Data":"93f1b8058553d8470f88ddb12317b750d0fb466cadf2900ffc773cf652b5f3b6"} Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.309574 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.333138 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.668328011 podStartE2EDuration="46.333118812s" podCreationTimestamp="2025-10-01 16:02:20 +0000 UTC" firstStartedPulling="2025-10-01 16:02:21.066236462 +0000 UTC m=+930.416876424" lastFinishedPulling="2025-10-01 16:03:05.731027263 +0000 UTC m=+975.081667225" observedRunningTime="2025-10-01 16:03:06.329160758 +0000 UTC m=+975.679800730" watchObservedRunningTime="2025-10-01 16:03:06.333118812 +0000 UTC m=+975.683758774" Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.612748 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7f469" Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.718077 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rwv8\" (UniqueName: \"kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8\") pod \"88217dd8-e3f4-4a77-8960-80980cda5e6a\" (UID: \"88217dd8-e3f4-4a77-8960-80980cda5e6a\") " Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.736194 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8" (OuterVolumeSpecName: "kube-api-access-8rwv8") pod "88217dd8-e3f4-4a77-8960-80980cda5e6a" (UID: "88217dd8-e3f4-4a77-8960-80980cda5e6a"). InnerVolumeSpecName "kube-api-access-8rwv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:06 crc kubenswrapper[4688]: I1001 16:03:06.819470 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rwv8\" (UniqueName: \"kubernetes.io/projected/88217dd8-e3f4-4a77-8960-80980cda5e6a-kube-api-access-8rwv8\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:07 crc kubenswrapper[4688]: I1001 16:03:07.318862 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7f469" event={"ID":"88217dd8-e3f4-4a77-8960-80980cda5e6a","Type":"ContainerDied","Data":"27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6"} Oct 01 16:03:07 crc kubenswrapper[4688]: I1001 16:03:07.318912 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a38ad9687839346b30fabf832a195328a5bf9cbc06f7018c84cba8839cd8f6" Oct 01 16:03:07 crc kubenswrapper[4688]: I1001 16:03:07.318969 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7f469" Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.332017 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5","Type":"ContainerStarted","Data":"b2b6b7f7ae0da18ce6e06e3f7e7c4c90ad06ef8c409d8ea5b3e27d43597ee551"} Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.332721 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.336692 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"33e5a114-5ed4-4df4-88ca-908340de2204","Type":"ContainerStarted","Data":"fe0e23f2ccc83eecd052ca534109114acfd4515a0087a05806d606be028389ee"} Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.336756 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"33e5a114-5ed4-4df4-88ca-908340de2204","Type":"ContainerStarted","Data":"c40f84abb413f3620ac43721f7a809ba6d69b4de7fd263dbe487fa989a4ab35e"} Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.336949 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.353804 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.6084673010000001 podStartE2EDuration="46.353789369s" podCreationTimestamp="2025-10-01 16:02:22 +0000 UTC" firstStartedPulling="2025-10-01 16:02:22.947557929 +0000 UTC m=+932.298197891" lastFinishedPulling="2025-10-01 16:03:07.692879997 +0000 UTC m=+977.043519959" observedRunningTime="2025-10-01 16:03:08.349259528 +0000 UTC m=+977.699899490" watchObservedRunningTime="2025-10-01 16:03:08.353789369 +0000 UTC m=+977.704429331" Oct 01 16:03:08 crc kubenswrapper[4688]: I1001 16:03:08.381797 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.260510547 podStartE2EDuration="8.381767134s" podCreationTimestamp="2025-10-01 16:03:00 +0000 UTC" firstStartedPulling="2025-10-01 16:03:01.519807939 +0000 UTC m=+970.870447891" lastFinishedPulling="2025-10-01 16:03:07.641064516 +0000 UTC m=+976.991704478" observedRunningTime="2025-10-01 16:03:08.369585453 +0000 UTC m=+977.720225425" watchObservedRunningTime="2025-10-01 16:03:08.381767134 +0000 UTC m=+977.732407126" Oct 01 16:03:09 crc kubenswrapper[4688]: I1001 16:03:09.349336 4688 generic.go:334] "Generic (PLEG): container finished" podID="81865451-e9a4-4f89-aa96-eb345800ffe3" containerID="40d79a868229c57fcf20ce7779ea879909827e8fb97f7f60731ac6de3a3cf871" exitCode=0 Oct 01 16:03:09 crc kubenswrapper[4688]: I1001 16:03:09.349398 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81865451-e9a4-4f89-aa96-eb345800ffe3","Type":"ContainerDied","Data":"40d79a868229c57fcf20ce7779ea879909827e8fb97f7f60731ac6de3a3cf871"} Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.222229 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lxw94"] Oct 01 16:03:10 crc kubenswrapper[4688]: E1001 16:03:10.223012 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88217dd8-e3f4-4a77-8960-80980cda5e6a" containerName="mariadb-database-create" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.223043 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="88217dd8-e3f4-4a77-8960-80980cda5e6a" containerName="mariadb-database-create" Oct 01 16:03:10 crc kubenswrapper[4688]: E1001 16:03:10.223061 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="dnsmasq-dns" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.223074 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="dnsmasq-dns" Oct 01 16:03:10 crc kubenswrapper[4688]: E1001 16:03:10.223111 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="init" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.223122 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="init" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.223394 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="81eed4a0-35bf-4aa0-b1c0-4208d39946d9" containerName="dnsmasq-dns" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.223437 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="88217dd8-e3f4-4a77-8960-80980cda5e6a" containerName="mariadb-database-create" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.224196 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.237997 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lxw94"] Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.359609 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"81865451-e9a4-4f89-aa96-eb345800ffe3","Type":"ContainerStarted","Data":"8f202a3f7617e12819ba9c2957608598819d2467b5e94ce22448d81d154eef71"} Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.378324 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7c4m\" (UniqueName: \"kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m\") pod \"keystone-db-create-lxw94\" (UID: \"d7dff43a-da4b-412d-a08e-eae2053bfb8c\") " pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.384914 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371984.469875 podStartE2EDuration="52.384899767s" podCreationTimestamp="2025-10-01 16:02:18 +0000 UTC" firstStartedPulling="2025-10-01 16:02:20.855859017 +0000 UTC m=+930.206498979" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:10.384452994 +0000 UTC m=+979.735092966" watchObservedRunningTime="2025-10-01 16:03:10.384899767 +0000 UTC m=+979.735539719" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.480285 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7c4m\" (UniqueName: \"kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m\") pod \"keystone-db-create-lxw94\" (UID: \"d7dff43a-da4b-412d-a08e-eae2053bfb8c\") " pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.519297 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7c4m\" (UniqueName: \"kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m\") pod \"keystone-db-create-lxw94\" (UID: \"d7dff43a-da4b-412d-a08e-eae2053bfb8c\") " pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.551993 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.631694 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.876993 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4grcg"] Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.878150 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4grcg" Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.884581 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4grcg"] Oct 01 16:03:10 crc kubenswrapper[4688]: I1001 16:03:10.987983 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prrwn\" (UniqueName: \"kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn\") pod \"glance-db-create-4grcg\" (UID: \"5227ef09-55af-48c6-a8ee-e6ff58bca51a\") " pod="openstack/glance-db-create-4grcg" Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.014207 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lxw94"] Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.092710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prrwn\" (UniqueName: \"kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn\") pod \"glance-db-create-4grcg\" (UID: \"5227ef09-55af-48c6-a8ee-e6ff58bca51a\") " pod="openstack/glance-db-create-4grcg" Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.113480 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prrwn\" (UniqueName: \"kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn\") pod \"glance-db-create-4grcg\" (UID: \"5227ef09-55af-48c6-a8ee-e6ff58bca51a\") " pod="openstack/glance-db-create-4grcg" Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.200477 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4grcg" Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.372547 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7dff43a-da4b-412d-a08e-eae2053bfb8c" containerID="80ec9296ac066bb97c864f35192c472ed3c0638029261e21b156f659d48a6c3f" exitCode=0 Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.372586 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxw94" event={"ID":"d7dff43a-da4b-412d-a08e-eae2053bfb8c","Type":"ContainerDied","Data":"80ec9296ac066bb97c864f35192c472ed3c0638029261e21b156f659d48a6c3f"} Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.372607 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxw94" event={"ID":"d7dff43a-da4b-412d-a08e-eae2053bfb8c","Type":"ContainerStarted","Data":"85fee8190f30ee98a0ef485286f0b221163b724b6d8a73d96130c1f3d4ed5373"} Oct 01 16:03:11 crc kubenswrapper[4688]: I1001 16:03:11.621469 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4grcg"] Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.393997 4688 generic.go:334] "Generic (PLEG): container finished" podID="5227ef09-55af-48c6-a8ee-e6ff58bca51a" containerID="47344ebf8c4c49c75b30b173377e7ba797c4caf38bd15fade03c857f475ee394" exitCode=0 Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.395845 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4grcg" event={"ID":"5227ef09-55af-48c6-a8ee-e6ff58bca51a","Type":"ContainerDied","Data":"47344ebf8c4c49c75b30b173377e7ba797c4caf38bd15fade03c857f475ee394"} Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.395896 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4grcg" event={"ID":"5227ef09-55af-48c6-a8ee-e6ff58bca51a","Type":"ContainerStarted","Data":"aa529ea8a356b7a6aecf30d0f0bcee95352482f6faf409b84e8536bbb8dfccca"} Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.479874 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.591414 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.593190 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.625264 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.727374 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.727422 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.727478 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgt8n\" (UniqueName: \"kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.727515 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.727569 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.829209 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.829298 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.829316 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.829365 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgt8n\" (UniqueName: \"kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.829388 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.830183 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.830675 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.831228 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.832042 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.856058 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgt8n\" (UniqueName: \"kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n\") pod \"dnsmasq-dns-b8fbc5445-77s2p\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.902456 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:12 crc kubenswrapper[4688]: I1001 16:03:12.919199 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.033222 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7c4m\" (UniqueName: \"kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m\") pod \"d7dff43a-da4b-412d-a08e-eae2053bfb8c\" (UID: \"d7dff43a-da4b-412d-a08e-eae2053bfb8c\") " Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.045356 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m" (OuterVolumeSpecName: "kube-api-access-v7c4m") pod "d7dff43a-da4b-412d-a08e-eae2053bfb8c" (UID: "d7dff43a-da4b-412d-a08e-eae2053bfb8c"). InnerVolumeSpecName "kube-api-access-v7c4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.134858 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7c4m\" (UniqueName: \"kubernetes.io/projected/d7dff43a-da4b-412d-a08e-eae2053bfb8c-kube-api-access-v7c4m\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.404984 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxw94" event={"ID":"d7dff43a-da4b-412d-a08e-eae2053bfb8c","Type":"ContainerDied","Data":"85fee8190f30ee98a0ef485286f0b221163b724b6d8a73d96130c1f3d4ed5373"} Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.405044 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85fee8190f30ee98a0ef485286f0b221163b724b6d8a73d96130c1f3d4ed5373" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.405004 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxw94" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.459910 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:03:13 crc kubenswrapper[4688]: W1001 16:03:13.465958 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e6228f_ade9_4baf_a01f_8d5e19e5836c.slice/crio-f3e8016c1d889b9dbf2e9590590ffed501c95788f88aea04a1dbd3d3a8981fcf WatchSource:0}: Error finding container f3e8016c1d889b9dbf2e9590590ffed501c95788f88aea04a1dbd3d3a8981fcf: Status 404 returned error can't find the container with id f3e8016c1d889b9dbf2e9590590ffed501c95788f88aea04a1dbd3d3a8981fcf Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.636633 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4grcg" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.687945 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:03:13 crc kubenswrapper[4688]: E1001 16:03:13.688335 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5227ef09-55af-48c6-a8ee-e6ff58bca51a" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.688357 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="5227ef09-55af-48c6-a8ee-e6ff58bca51a" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: E1001 16:03:13.688385 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7dff43a-da4b-412d-a08e-eae2053bfb8c" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.688393 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7dff43a-da4b-412d-a08e-eae2053bfb8c" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.688599 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7dff43a-da4b-412d-a08e-eae2053bfb8c" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.688616 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="5227ef09-55af-48c6-a8ee-e6ff58bca51a" containerName="mariadb-database-create" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.723789 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.723909 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.727038 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9vfw2" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.727228 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.727333 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.727461 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.751513 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prrwn\" (UniqueName: \"kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn\") pod \"5227ef09-55af-48c6-a8ee-e6ff58bca51a\" (UID: \"5227ef09-55af-48c6-a8ee-e6ff58bca51a\") " Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.761587 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn" (OuterVolumeSpecName: "kube-api-access-prrwn") pod "5227ef09-55af-48c6-a8ee-e6ff58bca51a" (UID: "5227ef09-55af-48c6-a8ee-e6ff58bca51a"). InnerVolumeSpecName "kube-api-access-prrwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.852726 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.852828 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqzc\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-kube-api-access-5nqzc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.852887 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-lock\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.852950 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-cache\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.853041 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.853683 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prrwn\" (UniqueName: \"kubernetes.io/projected/5227ef09-55af-48c6-a8ee-e6ff58bca51a-kube-api-access-prrwn\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955253 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-lock\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955438 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-cache\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955597 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955713 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955769 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqzc\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-kube-api-access-5nqzc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.955870 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-lock\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: E1001 16:03:13.955984 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:13 crc kubenswrapper[4688]: E1001 16:03:13.956009 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:13 crc kubenswrapper[4688]: E1001 16:03:13.956053 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:14.456034738 +0000 UTC m=+983.806674700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.956060 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.968378 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0cd57226-8163-4319-ad42-5a17df19881a-cache\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.977622 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqzc\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-kube-api-access-5nqzc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:13 crc kubenswrapper[4688]: I1001 16:03:13.980769 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.416826 4688 generic.go:334] "Generic (PLEG): container finished" podID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerID="9c0aa2322502dbe9c71c6829902f6b42366ba368d03602bca5cda9e653c79544" exitCode=0 Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.416914 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" event={"ID":"62e6228f-ade9-4baf-a01f-8d5e19e5836c","Type":"ContainerDied","Data":"9c0aa2322502dbe9c71c6829902f6b42366ba368d03602bca5cda9e653c79544"} Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.417083 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" event={"ID":"62e6228f-ade9-4baf-a01f-8d5e19e5836c","Type":"ContainerStarted","Data":"f3e8016c1d889b9dbf2e9590590ffed501c95788f88aea04a1dbd3d3a8981fcf"} Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.418703 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4grcg" event={"ID":"5227ef09-55af-48c6-a8ee-e6ff58bca51a","Type":"ContainerDied","Data":"aa529ea8a356b7a6aecf30d0f0bcee95352482f6faf409b84e8536bbb8dfccca"} Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.418738 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa529ea8a356b7a6aecf30d0f0bcee95352482f6faf409b84e8536bbb8dfccca" Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.419308 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4grcg" Oct 01 16:03:14 crc kubenswrapper[4688]: I1001 16:03:14.465783 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:14 crc kubenswrapper[4688]: E1001 16:03:14.466081 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:14 crc kubenswrapper[4688]: E1001 16:03:14.466101 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:14 crc kubenswrapper[4688]: E1001 16:03:14.466180 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:15.466160089 +0000 UTC m=+984.816800061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:15 crc kubenswrapper[4688]: I1001 16:03:15.428547 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" event={"ID":"62e6228f-ade9-4baf-a01f-8d5e19e5836c","Type":"ContainerStarted","Data":"e7b9aa1bdf9f1fb38248175acc41dd1be51e9764e75233db32aa665d91636bbc"} Oct 01 16:03:15 crc kubenswrapper[4688]: I1001 16:03:15.429012 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:15 crc kubenswrapper[4688]: I1001 16:03:15.452245 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" podStartSLOduration=3.45222438 podStartE2EDuration="3.45222438s" podCreationTimestamp="2025-10-01 16:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:15.450790999 +0000 UTC m=+984.801430981" watchObservedRunningTime="2025-10-01 16:03:15.45222438 +0000 UTC m=+984.802864352" Oct 01 16:03:15 crc kubenswrapper[4688]: I1001 16:03:15.482233 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:15 crc kubenswrapper[4688]: E1001 16:03:15.482440 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:15 crc kubenswrapper[4688]: E1001 16:03:15.482640 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:15 crc kubenswrapper[4688]: E1001 16:03:15.482741 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:17.482726648 +0000 UTC m=+986.833366610 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.514233 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:17 crc kubenswrapper[4688]: E1001 16:03:17.514427 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:17 crc kubenswrapper[4688]: E1001 16:03:17.514558 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:17 crc kubenswrapper[4688]: E1001 16:03:17.514611 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:21.514594208 +0000 UTC m=+990.865234260 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.640965 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-96jrl"] Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.641984 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.643981 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.644161 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.646110 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.681565 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-96jrl"] Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718012 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718059 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718078 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718108 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718172 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718201 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7ktp\" (UniqueName: \"kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.718296 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820500 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820647 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820685 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820714 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820770 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820814 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.820846 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7ktp\" (UniqueName: \"kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.821802 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.821929 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.822095 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.827619 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.827945 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.829350 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.844626 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7ktp\" (UniqueName: \"kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp\") pod \"swift-ring-rebalance-96jrl\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:17 crc kubenswrapper[4688]: I1001 16:03:17.959636 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:18 crc kubenswrapper[4688]: I1001 16:03:18.465732 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-96jrl"] Oct 01 16:03:19 crc kubenswrapper[4688]: I1001 16:03:19.461739 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-96jrl" event={"ID":"f47bc7a7-66e7-4994-985d-2ff7da317e55","Type":"ContainerStarted","Data":"81a476bce0aafbb95723c76245dd7415b1d704ef31373ae81f5d09dbb7659a0f"} Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.212334 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.212389 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.276793 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6bb3-account-create-8hlc5"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.277722 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.279719 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.283912 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.290717 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bb3-account-create-8hlc5"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.370326 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79zfs\" (UniqueName: \"kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs\") pod \"keystone-6bb3-account-create-8hlc5\" (UID: \"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a\") " pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.472576 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79zfs\" (UniqueName: \"kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs\") pod \"keystone-6bb3-account-create-8hlc5\" (UID: \"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a\") " pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.496545 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79zfs\" (UniqueName: \"kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs\") pod \"keystone-6bb3-account-create-8hlc5\" (UID: \"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a\") " pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.546244 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.607730 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.655262 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8661-account-create-q74m4"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.664695 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.665844 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8661-account-create-q74m4"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.667251 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.778288 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8djx\" (UniqueName: \"kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx\") pod \"placement-8661-account-create-q74m4\" (UID: \"c43636cd-b73f-4427-860a-7d9c40f0732a\") " pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.880190 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8djx\" (UniqueName: \"kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx\") pod \"placement-8661-account-create-q74m4\" (UID: \"c43636cd-b73f-4427-860a-7d9c40f0732a\") " pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.917253 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8djx\" (UniqueName: \"kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx\") pod \"placement-8661-account-create-q74m4\" (UID: \"c43636cd-b73f-4427-860a-7d9c40f0732a\") " pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.919949 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-570c-account-create-m78fq"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.921011 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.923220 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.931096 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-570c-account-create-m78fq"] Oct 01 16:03:20 crc kubenswrapper[4688]: I1001 16:03:20.986252 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.009371 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.086630 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv854\" (UniqueName: \"kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854\") pod \"glance-570c-account-create-m78fq\" (UID: \"05a1749d-a3be-4a23-888f-7ace91e261c2\") " pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.187845 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv854\" (UniqueName: \"kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854\") pod \"glance-570c-account-create-m78fq\" (UID: \"05a1749d-a3be-4a23-888f-7ace91e261c2\") " pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.211231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv854\" (UniqueName: \"kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854\") pod \"glance-570c-account-create-m78fq\" (UID: \"05a1749d-a3be-4a23-888f-7ace91e261c2\") " pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.258514 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:21 crc kubenswrapper[4688]: I1001 16:03:21.594175 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:21 crc kubenswrapper[4688]: E1001 16:03:21.594845 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:21 crc kubenswrapper[4688]: E1001 16:03:21.594872 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:21 crc kubenswrapper[4688]: E1001 16:03:21.594907 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:29.594893062 +0000 UTC m=+998.945533024 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.153499 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.197979 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ts722" podUID="e4559e8a-674a-427d-a9fb-145f94ff3b58" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:03:22 crc kubenswrapper[4688]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:03:22 crc kubenswrapper[4688]: > Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.734838 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bb3-account-create-8hlc5"] Oct 01 16:03:22 crc kubenswrapper[4688]: W1001 16:03:22.736284 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod568ac047_f5a3_4d4b_bf31_9a294e0ccf6a.slice/crio-3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd WatchSource:0}: Error finding container 3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd: Status 404 returned error can't find the container with id 3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.787241 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-570c-account-create-m78fq"] Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.873567 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8661-account-create-q74m4"] Oct 01 16:03:22 crc kubenswrapper[4688]: I1001 16:03:22.921828 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.013864 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.014139 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-nxw4l" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="dnsmasq-dns" containerID="cri-o://449ff1359abc109fda1da7fab8490b8aed89bbd63b67d91ae26cd761e242fa1f" gracePeriod=10 Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.495708 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-96jrl" event={"ID":"f47bc7a7-66e7-4994-985d-2ff7da317e55","Type":"ContainerStarted","Data":"97a1c01b83a4ee235893724bdf4253471d7fa2963776435393fca8cbbadf7c86"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.498033 4688 generic.go:334] "Generic (PLEG): container finished" podID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerID="449ff1359abc109fda1da7fab8490b8aed89bbd63b67d91ae26cd761e242fa1f" exitCode=0 Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.498080 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nxw4l" event={"ID":"79f808de-0fe9-4ed9-aa09-826434cd24bb","Type":"ContainerDied","Data":"449ff1359abc109fda1da7fab8490b8aed89bbd63b67d91ae26cd761e242fa1f"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.499093 4688 generic.go:334] "Generic (PLEG): container finished" podID="05a1749d-a3be-4a23-888f-7ace91e261c2" containerID="0e92bafaef8a011fbf78e883dce26a5cd1669424d6ca178c3ba37e6bd02e1bb9" exitCode=0 Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.499134 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-570c-account-create-m78fq" event={"ID":"05a1749d-a3be-4a23-888f-7ace91e261c2","Type":"ContainerDied","Data":"0e92bafaef8a011fbf78e883dce26a5cd1669424d6ca178c3ba37e6bd02e1bb9"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.499148 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-570c-account-create-m78fq" event={"ID":"05a1749d-a3be-4a23-888f-7ace91e261c2","Type":"ContainerStarted","Data":"9e248be601fd68f7b3a621f23a9d096538891f829e35c635826f6b7964083594"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.500132 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8661-account-create-q74m4" event={"ID":"c43636cd-b73f-4427-860a-7d9c40f0732a","Type":"ContainerStarted","Data":"1fd84cc04e29376ad465e8deea794a6aa5f8df429a4da30283ffba5038ec8ec5"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.500153 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8661-account-create-q74m4" event={"ID":"c43636cd-b73f-4427-860a-7d9c40f0732a","Type":"ContainerStarted","Data":"691035abb745f2a89203c5d8db1bd66c9eed993afd852f0aeb32aa30e9ca5524"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.501281 4688 generic.go:334] "Generic (PLEG): container finished" podID="568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" containerID="1765e397f8877587e83a24018aafd9a6d3fb2ef46b0dd4c821066f969842871a" exitCode=0 Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.501311 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bb3-account-create-8hlc5" event={"ID":"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a","Type":"ContainerDied","Data":"1765e397f8877587e83a24018aafd9a6d3fb2ef46b0dd4c821066f969842871a"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.501328 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bb3-account-create-8hlc5" event={"ID":"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a","Type":"ContainerStarted","Data":"3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd"} Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.524711 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-96jrl" podStartSLOduration=2.699026887 podStartE2EDuration="6.524688994s" podCreationTimestamp="2025-10-01 16:03:17 +0000 UTC" firstStartedPulling="2025-10-01 16:03:18.46845968 +0000 UTC m=+987.819099642" lastFinishedPulling="2025-10-01 16:03:22.294121787 +0000 UTC m=+991.644761749" observedRunningTime="2025-10-01 16:03:23.517885318 +0000 UTC m=+992.868525300" watchObservedRunningTime="2025-10-01 16:03:23.524688994 +0000 UTC m=+992.875328956" Oct 01 16:03:23 crc kubenswrapper[4688]: I1001 16:03:23.936252 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.042300 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config\") pod \"79f808de-0fe9-4ed9-aa09-826434cd24bb\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.042357 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb\") pod \"79f808de-0fe9-4ed9-aa09-826434cd24bb\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.042378 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc\") pod \"79f808de-0fe9-4ed9-aa09-826434cd24bb\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.042473 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6rht\" (UniqueName: \"kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht\") pod \"79f808de-0fe9-4ed9-aa09-826434cd24bb\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.043369 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb\") pod \"79f808de-0fe9-4ed9-aa09-826434cd24bb\" (UID: \"79f808de-0fe9-4ed9-aa09-826434cd24bb\") " Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.060050 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht" (OuterVolumeSpecName: "kube-api-access-k6rht") pod "79f808de-0fe9-4ed9-aa09-826434cd24bb" (UID: "79f808de-0fe9-4ed9-aa09-826434cd24bb"). InnerVolumeSpecName "kube-api-access-k6rht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.088024 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79f808de-0fe9-4ed9-aa09-826434cd24bb" (UID: "79f808de-0fe9-4ed9-aa09-826434cd24bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.088724 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "79f808de-0fe9-4ed9-aa09-826434cd24bb" (UID: "79f808de-0fe9-4ed9-aa09-826434cd24bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.089188 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "79f808de-0fe9-4ed9-aa09-826434cd24bb" (UID: "79f808de-0fe9-4ed9-aa09-826434cd24bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.091092 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config" (OuterVolumeSpecName: "config") pod "79f808de-0fe9-4ed9-aa09-826434cd24bb" (UID: "79f808de-0fe9-4ed9-aa09-826434cd24bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.145272 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.145307 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.145318 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.145329 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6rht\" (UniqueName: \"kubernetes.io/projected/79f808de-0fe9-4ed9-aa09-826434cd24bb-kube-api-access-k6rht\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.145339 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79f808de-0fe9-4ed9-aa09-826434cd24bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.513493 4688 generic.go:334] "Generic (PLEG): container finished" podID="c43636cd-b73f-4427-860a-7d9c40f0732a" containerID="1fd84cc04e29376ad465e8deea794a6aa5f8df429a4da30283ffba5038ec8ec5" exitCode=0 Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.513635 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8661-account-create-q74m4" event={"ID":"c43636cd-b73f-4427-860a-7d9c40f0732a","Type":"ContainerDied","Data":"1fd84cc04e29376ad465e8deea794a6aa5f8df429a4da30283ffba5038ec8ec5"} Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.517677 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nxw4l" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.518184 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nxw4l" event={"ID":"79f808de-0fe9-4ed9-aa09-826434cd24bb","Type":"ContainerDied","Data":"2896ecd3183ef7f6423ae8b7d4cb73cecd5ecf5419508d70dfc524f825fcf31d"} Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.518314 4688 scope.go:117] "RemoveContainer" containerID="449ff1359abc109fda1da7fab8490b8aed89bbd63b67d91ae26cd761e242fa1f" Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.573747 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.583493 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nxw4l"] Oct 01 16:03:24 crc kubenswrapper[4688]: I1001 16:03:24.588972 4688 scope.go:117] "RemoveContainer" containerID="600adccb09146af7cf9c0aca8bb3648a97d7928cf37a920becba6f8c7919d0fd" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.008696 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.019624 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.031811 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.163409 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8djx\" (UniqueName: \"kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx\") pod \"c43636cd-b73f-4427-860a-7d9c40f0732a\" (UID: \"c43636cd-b73f-4427-860a-7d9c40f0732a\") " Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.163512 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79zfs\" (UniqueName: \"kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs\") pod \"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a\" (UID: \"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a\") " Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.163575 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv854\" (UniqueName: \"kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854\") pod \"05a1749d-a3be-4a23-888f-7ace91e261c2\" (UID: \"05a1749d-a3be-4a23-888f-7ace91e261c2\") " Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.170000 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx" (OuterVolumeSpecName: "kube-api-access-m8djx") pod "c43636cd-b73f-4427-860a-7d9c40f0732a" (UID: "c43636cd-b73f-4427-860a-7d9c40f0732a"). InnerVolumeSpecName "kube-api-access-m8djx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.175786 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs" (OuterVolumeSpecName: "kube-api-access-79zfs") pod "568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" (UID: "568ac047-f5a3-4d4b-bf31-9a294e0ccf6a"). InnerVolumeSpecName "kube-api-access-79zfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.176457 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854" (OuterVolumeSpecName: "kube-api-access-bv854") pod "05a1749d-a3be-4a23-888f-7ace91e261c2" (UID: "05a1749d-a3be-4a23-888f-7ace91e261c2"). InnerVolumeSpecName "kube-api-access-bv854". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.265057 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8djx\" (UniqueName: \"kubernetes.io/projected/c43636cd-b73f-4427-860a-7d9c40f0732a-kube-api-access-m8djx\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.265091 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79zfs\" (UniqueName: \"kubernetes.io/projected/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a-kube-api-access-79zfs\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.265101 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv854\" (UniqueName: \"kubernetes.io/projected/05a1749d-a3be-4a23-888f-7ace91e261c2-kube-api-access-bv854\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.389873 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" path="/var/lib/kubelet/pods/79f808de-0fe9-4ed9-aa09-826434cd24bb/volumes" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.532403 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-570c-account-create-m78fq" event={"ID":"05a1749d-a3be-4a23-888f-7ace91e261c2","Type":"ContainerDied","Data":"9e248be601fd68f7b3a621f23a9d096538891f829e35c635826f6b7964083594"} Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.532444 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e248be601fd68f7b3a621f23a9d096538891f829e35c635826f6b7964083594" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.532442 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-570c-account-create-m78fq" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.536037 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8661-account-create-q74m4" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.536233 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8661-account-create-q74m4" event={"ID":"c43636cd-b73f-4427-860a-7d9c40f0732a","Type":"ContainerDied","Data":"691035abb745f2a89203c5d8db1bd66c9eed993afd852f0aeb32aa30e9ca5524"} Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.536278 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="691035abb745f2a89203c5d8db1bd66c9eed993afd852f0aeb32aa30e9ca5524" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.539949 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bb3-account-create-8hlc5" event={"ID":"568ac047-f5a3-4d4b-bf31-9a294e0ccf6a","Type":"ContainerDied","Data":"3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd"} Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.539973 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a1533b7e7c810fe1363268d6a9482e5ee5efcd3f60b987f3d16ade3704cfcbd" Oct 01 16:03:25 crc kubenswrapper[4688]: I1001 16:03:25.540009 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bb3-account-create-8hlc5" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.137128 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9wpf7" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.169566 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ts722" podUID="e4559e8a-674a-427d-a9fb-145f94ff3b58" containerName="ovn-controller" probeResult="failure" output=< Oct 01 16:03:27 crc kubenswrapper[4688]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 16:03:27 crc kubenswrapper[4688]: > Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.344905 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ts722-config-6bg62"] Oct 01 16:03:27 crc kubenswrapper[4688]: E1001 16:03:27.345326 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345348 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: E1001 16:03:27.345363 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="init" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345370 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="init" Oct 01 16:03:27 crc kubenswrapper[4688]: E1001 16:03:27.345387 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="dnsmasq-dns" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345395 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="dnsmasq-dns" Oct 01 16:03:27 crc kubenswrapper[4688]: E1001 16:03:27.345415 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c43636cd-b73f-4427-860a-7d9c40f0732a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345424 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="c43636cd-b73f-4427-860a-7d9c40f0732a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: E1001 16:03:27.345439 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05a1749d-a3be-4a23-888f-7ace91e261c2" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345446 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="05a1749d-a3be-4a23-888f-7ace91e261c2" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345651 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345674 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="05a1749d-a3be-4a23-888f-7ace91e261c2" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345686 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="c43636cd-b73f-4427-860a-7d9c40f0732a" containerName="mariadb-account-create" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.345704 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f808de-0fe9-4ed9-aa09-826434cd24bb" containerName="dnsmasq-dns" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.346291 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.348671 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.371440 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ts722-config-6bg62"] Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509236 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509307 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509344 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509364 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509396 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crg6r\" (UniqueName: \"kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.509608 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611314 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611474 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611515 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611574 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611597 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611634 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crg6r\" (UniqueName: \"kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611799 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611816 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.611884 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.612368 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.613567 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.639435 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crg6r\" (UniqueName: \"kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r\") pod \"ovn-controller-ts722-config-6bg62\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:27 crc kubenswrapper[4688]: I1001 16:03:27.662533 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:28 crc kubenswrapper[4688]: I1001 16:03:28.157215 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ts722-config-6bg62"] Oct 01 16:03:28 crc kubenswrapper[4688]: I1001 16:03:28.564215 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722-config-6bg62" event={"ID":"b143244f-fb8a-4678-8722-d80070557c42","Type":"ContainerStarted","Data":"56373d3e4abb22326b033e7c777f0942a0224d5d62b4c5463e1734dd83e33749"} Oct 01 16:03:28 crc kubenswrapper[4688]: I1001 16:03:28.564266 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722-config-6bg62" event={"ID":"b143244f-fb8a-4678-8722-d80070557c42","Type":"ContainerStarted","Data":"59356a4717ef8ab6c986db9a7e61819a1d5bdd63307a567a70e34cddc20593b9"} Oct 01 16:03:28 crc kubenswrapper[4688]: I1001 16:03:28.587007 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ts722-config-6bg62" podStartSLOduration=1.586986512 podStartE2EDuration="1.586986512s" podCreationTimestamp="2025-10-01 16:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:28.580946258 +0000 UTC m=+997.931586220" watchObservedRunningTime="2025-10-01 16:03:28.586986512 +0000 UTC m=+997.937626474" Oct 01 16:03:29 crc kubenswrapper[4688]: I1001 16:03:29.572772 4688 generic.go:334] "Generic (PLEG): container finished" podID="b143244f-fb8a-4678-8722-d80070557c42" containerID="56373d3e4abb22326b033e7c777f0942a0224d5d62b4c5463e1734dd83e33749" exitCode=0 Oct 01 16:03:29 crc kubenswrapper[4688]: I1001 16:03:29.572869 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722-config-6bg62" event={"ID":"b143244f-fb8a-4678-8722-d80070557c42","Type":"ContainerDied","Data":"56373d3e4abb22326b033e7c777f0942a0224d5d62b4c5463e1734dd83e33749"} Oct 01 16:03:29 crc kubenswrapper[4688]: I1001 16:03:29.654164 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:29 crc kubenswrapper[4688]: E1001 16:03:29.655141 4688 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 16:03:29 crc kubenswrapper[4688]: E1001 16:03:29.655421 4688 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 16:03:29 crc kubenswrapper[4688]: E1001 16:03:29.655706 4688 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift podName:0cd57226-8163-4319-ad42-5a17df19881a nodeName:}" failed. No retries permitted until 2025-10-01 16:03:45.6556688 +0000 UTC m=+1015.006308792 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift") pod "swift-storage-0" (UID: "0cd57226-8163-4319-ad42-5a17df19881a") : configmap "swift-ring-files" not found Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.587966 4688 generic.go:334] "Generic (PLEG): container finished" podID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerID="f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567" exitCode=0 Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.588060 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerDied","Data":"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567"} Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.592401 4688 generic.go:334] "Generic (PLEG): container finished" podID="f47bc7a7-66e7-4994-985d-2ff7da317e55" containerID="97a1c01b83a4ee235893724bdf4253471d7fa2963776435393fca8cbbadf7c86" exitCode=0 Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.592466 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-96jrl" event={"ID":"f47bc7a7-66e7-4994-985d-2ff7da317e55","Type":"ContainerDied","Data":"97a1c01b83a4ee235893724bdf4253471d7fa2963776435393fca8cbbadf7c86"} Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.595711 4688 generic.go:334] "Generic (PLEG): container finished" podID="0372e042-2565-4498-8884-452f6e0bd0f1" containerID="b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d" exitCode=0 Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.595786 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerDied","Data":"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d"} Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.914999 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985607 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985746 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985835 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985927 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crg6r\" (UniqueName: \"kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985977 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.986004 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run\") pod \"b143244f-fb8a-4678-8722-d80070557c42\" (UID: \"b143244f-fb8a-4678-8722-d80070557c42\") " Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.985691 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.986781 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.987140 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.987171 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts" (OuterVolumeSpecName: "scripts") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.987722 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run" (OuterVolumeSpecName: "var-run") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.991114 4688 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.991153 4688 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.991164 4688 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.991175 4688 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b143244f-fb8a-4678-8722-d80070557c42-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.991201 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b143244f-fb8a-4678-8722-d80070557c42-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:30 crc kubenswrapper[4688]: I1001 16:03:30.992023 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r" (OuterVolumeSpecName: "kube-api-access-crg6r") pod "b143244f-fb8a-4678-8722-d80070557c42" (UID: "b143244f-fb8a-4678-8722-d80070557c42"). InnerVolumeSpecName "kube-api-access-crg6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.092508 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crg6r\" (UniqueName: \"kubernetes.io/projected/b143244f-fb8a-4678-8722-d80070557c42-kube-api-access-crg6r\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.164864 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-l5md4"] Oct 01 16:03:31 crc kubenswrapper[4688]: E1001 16:03:31.181163 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b143244f-fb8a-4678-8722-d80070557c42" containerName="ovn-config" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.181433 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b143244f-fb8a-4678-8722-d80070557c42" containerName="ovn-config" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.187817 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b143244f-fb8a-4678-8722-d80070557c42" containerName="ovn-config" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.188955 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.192095 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.192509 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pbr6z" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.226376 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l5md4"] Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.295465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.295639 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.295668 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6br86\" (UniqueName: \"kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.295701 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.396970 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.397260 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6br86\" (UniqueName: \"kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.397349 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.397435 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.400682 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.403730 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.404418 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.415401 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6br86\" (UniqueName: \"kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86\") pod \"glance-db-sync-l5md4\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.542267 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l5md4" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.606628 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerStarted","Data":"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd"} Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.607059 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.621972 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerStarted","Data":"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49"} Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.622206 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.624739 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ts722-config-6bg62" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.625446 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ts722-config-6bg62" event={"ID":"b143244f-fb8a-4678-8722-d80070557c42","Type":"ContainerDied","Data":"59356a4717ef8ab6c986db9a7e61819a1d5bdd63307a567a70e34cddc20593b9"} Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.625476 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59356a4717ef8ab6c986db9a7e61819a1d5bdd63307a567a70e34cddc20593b9" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.660173 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.879935976 podStartE2EDuration="1m16.660155622s" podCreationTimestamp="2025-10-01 16:02:15 +0000 UTC" firstStartedPulling="2025-10-01 16:02:17.686232732 +0000 UTC m=+927.036872684" lastFinishedPulling="2025-10-01 16:02:56.466452368 +0000 UTC m=+965.817092330" observedRunningTime="2025-10-01 16:03:31.641846475 +0000 UTC m=+1000.992486437" watchObservedRunningTime="2025-10-01 16:03:31.660155622 +0000 UTC m=+1001.010795584" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.671765 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.427742762 podStartE2EDuration="1m15.671745816s" podCreationTimestamp="2025-10-01 16:02:16 +0000 UTC" firstStartedPulling="2025-10-01 16:02:18.045966965 +0000 UTC m=+927.396606927" lastFinishedPulling="2025-10-01 16:02:56.289970019 +0000 UTC m=+965.640609981" observedRunningTime="2025-10-01 16:03:31.668511863 +0000 UTC m=+1001.019151825" watchObservedRunningTime="2025-10-01 16:03:31.671745816 +0000 UTC m=+1001.022385778" Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.755851 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ts722-config-6bg62"] Oct 01 16:03:31 crc kubenswrapper[4688]: I1001 16:03:31.763464 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ts722-config-6bg62"] Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.097877 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.184504 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ts722" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211157 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211253 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211326 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7ktp\" (UniqueName: \"kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211388 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211411 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211468 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.211619 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf\") pod \"f47bc7a7-66e7-4994-985d-2ff7da317e55\" (UID: \"f47bc7a7-66e7-4994-985d-2ff7da317e55\") " Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.212018 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.212326 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.212840 4688 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f47bc7a7-66e7-4994-985d-2ff7da317e55-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.212858 4688 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.223690 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp" (OuterVolumeSpecName: "kube-api-access-d7ktp") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "kube-api-access-d7ktp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.246237 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l5md4"] Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.249502 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.260877 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.264827 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.277170 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts" (OuterVolumeSpecName: "scripts") pod "f47bc7a7-66e7-4994-985d-2ff7da317e55" (UID: "f47bc7a7-66e7-4994-985d-2ff7da317e55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.314511 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7ktp\" (UniqueName: \"kubernetes.io/projected/f47bc7a7-66e7-4994-985d-2ff7da317e55-kube-api-access-d7ktp\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.314570 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.314583 4688 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.314599 4688 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f47bc7a7-66e7-4994-985d-2ff7da317e55-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.314613 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f47bc7a7-66e7-4994-985d-2ff7da317e55-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.632552 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-96jrl" event={"ID":"f47bc7a7-66e7-4994-985d-2ff7da317e55","Type":"ContainerDied","Data":"81a476bce0aafbb95723c76245dd7415b1d704ef31373ae81f5d09dbb7659a0f"} Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.632587 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a476bce0aafbb95723c76245dd7415b1d704ef31373ae81f5d09dbb7659a0f" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.632563 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-96jrl" Oct 01 16:03:32 crc kubenswrapper[4688]: I1001 16:03:32.634301 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l5md4" event={"ID":"ad6973b9-8d39-441f-ae39-67b632e06bcf","Type":"ContainerStarted","Data":"d285a82a34e0dad93f3561240365b102316c358cf0455a360f41098981c97420"} Oct 01 16:03:33 crc kubenswrapper[4688]: I1001 16:03:33.398387 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b143244f-fb8a-4678-8722-d80070557c42" path="/var/lib/kubelet/pods/b143244f-fb8a-4678-8722-d80070557c42/volumes" Oct 01 16:03:45 crc kubenswrapper[4688]: I1001 16:03:45.677972 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:45 crc kubenswrapper[4688]: I1001 16:03:45.691733 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0cd57226-8163-4319-ad42-5a17df19881a-etc-swift\") pod \"swift-storage-0\" (UID: \"0cd57226-8163-4319-ad42-5a17df19881a\") " pod="openstack/swift-storage-0" Oct 01 16:03:45 crc kubenswrapper[4688]: I1001 16:03:45.872567 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 16:03:46 crc kubenswrapper[4688]: I1001 16:03:46.402971 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 16:03:46 crc kubenswrapper[4688]: I1001 16:03:46.751767 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l5md4" event={"ID":"ad6973b9-8d39-441f-ae39-67b632e06bcf","Type":"ContainerStarted","Data":"46fcbc2ddeab50f23f49d9c431eb49e6459975bf3203d72e65912f21aa081c8a"} Oct 01 16:03:46 crc kubenswrapper[4688]: I1001 16:03:46.753357 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"e85d055188c0ff2dd81fa4c8adbd7d1b01bddadcc24dab9cbff2ecca8d5eceeb"} Oct 01 16:03:46 crc kubenswrapper[4688]: I1001 16:03:46.770382 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-l5md4" podStartSLOduration=2.223484884 podStartE2EDuration="15.770364909s" podCreationTimestamp="2025-10-01 16:03:31 +0000 UTC" firstStartedPulling="2025-10-01 16:03:32.268369356 +0000 UTC m=+1001.619009318" lastFinishedPulling="2025-10-01 16:03:45.815249371 +0000 UTC m=+1015.165889343" observedRunningTime="2025-10-01 16:03:46.765336855 +0000 UTC m=+1016.115976827" watchObservedRunningTime="2025-10-01 16:03:46.770364909 +0000 UTC m=+1016.121004871" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.145482 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.518718 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.602344 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-dq424"] Oct 01 16:03:47 crc kubenswrapper[4688]: E1001 16:03:47.602647 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47bc7a7-66e7-4994-985d-2ff7da317e55" containerName="swift-ring-rebalance" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.602662 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47bc7a7-66e7-4994-985d-2ff7da317e55" containerName="swift-ring-rebalance" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.602814 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f47bc7a7-66e7-4994-985d-2ff7da317e55" containerName="swift-ring-rebalance" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.603281 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dq424" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.618427 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdrv2\" (UniqueName: \"kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2\") pod \"cinder-db-create-dq424\" (UID: \"bb2febea-70d3-47ed-9ee2-1daf35d13f86\") " pod="openstack/cinder-db-create-dq424" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.625950 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dq424"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.685312 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bqz97"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.692104 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.698326 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bqz97"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.719465 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdrv2\" (UniqueName: \"kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2\") pod \"cinder-db-create-dq424\" (UID: \"bb2febea-70d3-47ed-9ee2-1daf35d13f86\") " pod="openstack/cinder-db-create-dq424" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.757234 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdrv2\" (UniqueName: \"kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2\") pod \"cinder-db-create-dq424\" (UID: \"bb2febea-70d3-47ed-9ee2-1daf35d13f86\") " pod="openstack/cinder-db-create-dq424" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.821069 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5brx\" (UniqueName: \"kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx\") pod \"barbican-db-create-bqz97\" (UID: \"efdacba2-e02a-4105-bd69-8791577b722c\") " pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.884747 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-nhpdd"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.885821 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.897699 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nhpdd"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.924162 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5brx\" (UniqueName: \"kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx\") pod \"barbican-db-create-bqz97\" (UID: \"efdacba2-e02a-4105-bd69-8791577b722c\") " pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.927417 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dq424" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.972263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5brx\" (UniqueName: \"kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx\") pod \"barbican-db-create-bqz97\" (UID: \"efdacba2-e02a-4105-bd69-8791577b722c\") " pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.979470 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bk86d"] Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.980721 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.985949 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcpfl" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.986150 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.986261 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:03:47 crc kubenswrapper[4688]: I1001 16:03:47.997336 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.010956 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bk86d"] Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.026136 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhjrf\" (UniqueName: \"kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf\") pod \"neutron-db-create-nhpdd\" (UID: \"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e\") " pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.026262 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.127699 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5wr4\" (UniqueName: \"kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.127756 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.127777 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.127872 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjrf\" (UniqueName: \"kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf\") pod \"neutron-db-create-nhpdd\" (UID: \"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e\") " pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.145353 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjrf\" (UniqueName: \"kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf\") pod \"neutron-db-create-nhpdd\" (UID: \"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e\") " pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.213881 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.228991 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5wr4\" (UniqueName: \"kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.229066 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.229097 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.246280 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.247080 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5wr4\" (UniqueName: \"kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.251064 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data\") pod \"keystone-db-sync-bk86d\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.313474 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bk86d" Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.771420 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"fb94884f31d59b7fdde7f4a77933f7c3b48a348a4ff693e38727988c6effb248"} Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.860825 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bqz97"] Oct 01 16:03:48 crc kubenswrapper[4688]: I1001 16:03:48.966181 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bk86d"] Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:48.985450 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nhpdd"] Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.045070 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dq424"] Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.785191 4688 generic.go:334] "Generic (PLEG): container finished" podID="efdacba2-e02a-4105-bd69-8791577b722c" containerID="4246ab03e0c4431300f6451df412d3c177eef48fd8b985b6b34fef0d3019c9ec" exitCode=0 Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.785755 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bqz97" event={"ID":"efdacba2-e02a-4105-bd69-8791577b722c","Type":"ContainerDied","Data":"4246ab03e0c4431300f6451df412d3c177eef48fd8b985b6b34fef0d3019c9ec"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.786302 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bqz97" event={"ID":"efdacba2-e02a-4105-bd69-8791577b722c","Type":"ContainerStarted","Data":"6165bc8adc45760d259ffd6768c982c3b405e268cecf3124113db41cdf79d5cf"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.790966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"b8b214ae15b184d7eac5e618f489579c22a22ec31d81e2304f0ce72639993440"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.791022 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"5ff58d5fa562f6c986f7a3a80df10f0efe4b574f57beb0c170675a07a13f0758"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.791033 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"6d20317e46f3b3039e8419e8adac1c10c7f802c513d4fc0c2d771fa200970454"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.800041 4688 generic.go:334] "Generic (PLEG): container finished" podID="7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" containerID="a1b0b43b9fdaf60a9ee4a24a49a01d3ba5423fea699d88c2dbf7c00bef2b06f7" exitCode=0 Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.800237 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nhpdd" event={"ID":"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e","Type":"ContainerDied","Data":"a1b0b43b9fdaf60a9ee4a24a49a01d3ba5423fea699d88c2dbf7c00bef2b06f7"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.800266 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nhpdd" event={"ID":"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e","Type":"ContainerStarted","Data":"ddebf58dcbbd4927ab918abad8782a1fe53d934681f2a23379faea2f901bd6b0"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.806276 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bk86d" event={"ID":"e322d157-d4b5-4b2c-92a6-141481c2b358","Type":"ContainerStarted","Data":"cd94deb706654f4662eddbe2ffa353192586e9090dd2aa70c157756e3d1d50fb"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.814210 4688 generic.go:334] "Generic (PLEG): container finished" podID="bb2febea-70d3-47ed-9ee2-1daf35d13f86" containerID="67e47d68212350a1d0a70a8831ae241af4c59011b5c34123b1e0c6d0d5619793" exitCode=0 Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.814507 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dq424" event={"ID":"bb2febea-70d3-47ed-9ee2-1daf35d13f86","Type":"ContainerDied","Data":"67e47d68212350a1d0a70a8831ae241af4c59011b5c34123b1e0c6d0d5619793"} Oct 01 16:03:49 crc kubenswrapper[4688]: I1001 16:03:49.814626 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dq424" event={"ID":"bb2febea-70d3-47ed-9ee2-1daf35d13f86","Type":"ContainerStarted","Data":"2189140b7af0e11028efc4ac9acfe624265664fb8c063ba45daff95cd33e642c"} Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.489230 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dq424" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.631469 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdrv2\" (UniqueName: \"kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2\") pod \"bb2febea-70d3-47ed-9ee2-1daf35d13f86\" (UID: \"bb2febea-70d3-47ed-9ee2-1daf35d13f86\") " Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.637407 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2" (OuterVolumeSpecName: "kube-api-access-rdrv2") pod "bb2febea-70d3-47ed-9ee2-1daf35d13f86" (UID: "bb2febea-70d3-47ed-9ee2-1daf35d13f86"). InnerVolumeSpecName "kube-api-access-rdrv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.694723 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.703105 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.737900 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdrv2\" (UniqueName: \"kubernetes.io/projected/bb2febea-70d3-47ed-9ee2-1daf35d13f86-kube-api-access-rdrv2\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.830130 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bqz97" event={"ID":"efdacba2-e02a-4105-bd69-8791577b722c","Type":"ContainerDied","Data":"6165bc8adc45760d259ffd6768c982c3b405e268cecf3124113db41cdf79d5cf"} Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.830167 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6165bc8adc45760d259ffd6768c982c3b405e268cecf3124113db41cdf79d5cf" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.830215 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bqz97" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.831869 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nhpdd" event={"ID":"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e","Type":"ContainerDied","Data":"ddebf58dcbbd4927ab918abad8782a1fe53d934681f2a23379faea2f901bd6b0"} Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.831898 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddebf58dcbbd4927ab918abad8782a1fe53d934681f2a23379faea2f901bd6b0" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.831943 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nhpdd" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.835447 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dq424" event={"ID":"bb2febea-70d3-47ed-9ee2-1daf35d13f86","Type":"ContainerDied","Data":"2189140b7af0e11028efc4ac9acfe624265664fb8c063ba45daff95cd33e642c"} Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.835476 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2189140b7af0e11028efc4ac9acfe624265664fb8c063ba45daff95cd33e642c" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.835705 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dq424" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.838436 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5brx\" (UniqueName: \"kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx\") pod \"efdacba2-e02a-4105-bd69-8791577b722c\" (UID: \"efdacba2-e02a-4105-bd69-8791577b722c\") " Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.838723 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhjrf\" (UniqueName: \"kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf\") pod \"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e\" (UID: \"7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e\") " Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.842178 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf" (OuterVolumeSpecName: "kube-api-access-vhjrf") pod "7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" (UID: "7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e"). InnerVolumeSpecName "kube-api-access-vhjrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.843772 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx" (OuterVolumeSpecName: "kube-api-access-r5brx") pod "efdacba2-e02a-4105-bd69-8791577b722c" (UID: "efdacba2-e02a-4105-bd69-8791577b722c"). InnerVolumeSpecName "kube-api-access-r5brx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.940458 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhjrf\" (UniqueName: \"kubernetes.io/projected/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e-kube-api-access-vhjrf\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:51 crc kubenswrapper[4688]: I1001 16:03:51.940494 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5brx\" (UniqueName: \"kubernetes.io/projected/efdacba2-e02a-4105-bd69-8791577b722c-kube-api-access-r5brx\") on node \"crc\" DevicePath \"\"" Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.925102 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"26deee3f474180c6ae86c7617cfaa54f7a524a0986657650a8f17b191c232788"} Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.925667 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"2fdbf31a5cc3c072a05871f4c369827ca77a87747196da30478ebe4404b518c0"} Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.925679 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"5b8c3ea4e1dbb7f88bcfec0bbdd0848c610a6d637b0a6ddd1536ac253d7113f0"} Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.925688 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"616e81761aa7d612dea474ea51e4b936a0bd384ea444e3e67271b565974213e7"} Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.931987 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bk86d" event={"ID":"e322d157-d4b5-4b2c-92a6-141481c2b358","Type":"ContainerStarted","Data":"de3e989d53a2da480bed2e6dae2272057e53c483fce3ba61e00724c73afff06b"} Oct 01 16:03:55 crc kubenswrapper[4688]: I1001 16:03:55.958714 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bk86d" podStartSLOduration=2.689473155 podStartE2EDuration="8.9586976s" podCreationTimestamp="2025-10-01 16:03:47 +0000 UTC" firstStartedPulling="2025-10-01 16:03:48.999753304 +0000 UTC m=+1018.350393266" lastFinishedPulling="2025-10-01 16:03:55.268977749 +0000 UTC m=+1024.619617711" observedRunningTime="2025-10-01 16:03:55.956838996 +0000 UTC m=+1025.307478958" watchObservedRunningTime="2025-10-01 16:03:55.9586976 +0000 UTC m=+1025.309337562" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.609699 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-52dc-account-create-jqvjq"] Oct 01 16:03:57 crc kubenswrapper[4688]: E1001 16:03:57.610620 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610637 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: E1001 16:03:57.610652 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb2febea-70d3-47ed-9ee2-1daf35d13f86" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610660 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb2febea-70d3-47ed-9ee2-1daf35d13f86" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: E1001 16:03:57.610681 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdacba2-e02a-4105-bd69-8791577b722c" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610689 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdacba2-e02a-4105-bd69-8791577b722c" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610895 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb2febea-70d3-47ed-9ee2-1daf35d13f86" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610936 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.610955 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdacba2-e02a-4105-bd69-8791577b722c" containerName="mariadb-database-create" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.611476 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.616381 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.627349 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-52dc-account-create-jqvjq"] Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.657110 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k95cr\" (UniqueName: \"kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr\") pod \"cinder-52dc-account-create-jqvjq\" (UID: \"59228674-5e5c-426f-9e2d-e82c5760b356\") " pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.758040 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k95cr\" (UniqueName: \"kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr\") pod \"cinder-52dc-account-create-jqvjq\" (UID: \"59228674-5e5c-426f-9e2d-e82c5760b356\") " pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.781289 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k95cr\" (UniqueName: \"kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr\") pod \"cinder-52dc-account-create-jqvjq\" (UID: \"59228674-5e5c-426f-9e2d-e82c5760b356\") " pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.818714 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d7e0-account-create-58kmq"] Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.822779 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.826694 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.830158 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d7e0-account-create-58kmq"] Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.859165 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c28ql\" (UniqueName: \"kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql\") pod \"barbican-d7e0-account-create-58kmq\" (UID: \"f8c4d47a-f834-4952-a414-34fb8df2a1ef\") " pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.936053 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.961207 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c28ql\" (UniqueName: \"kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql\") pod \"barbican-d7e0-account-create-58kmq\" (UID: \"f8c4d47a-f834-4952-a414-34fb8df2a1ef\") " pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:03:57 crc kubenswrapper[4688]: I1001 16:03:57.983329 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c28ql\" (UniqueName: \"kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql\") pod \"barbican-d7e0-account-create-58kmq\" (UID: \"f8c4d47a-f834-4952-a414-34fb8df2a1ef\") " pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.016992 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-058d-account-create-fzsds"] Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.018732 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.025860 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-058d-account-create-fzsds"] Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.026128 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.064093 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb75\" (UniqueName: \"kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75\") pod \"neutron-058d-account-create-fzsds\" (UID: \"ce461b22-96d9-42ee-950d-b001054c5c60\") " pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.143715 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.165967 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb75\" (UniqueName: \"kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75\") pod \"neutron-058d-account-create-fzsds\" (UID: \"ce461b22-96d9-42ee-950d-b001054c5c60\") " pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.190117 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb75\" (UniqueName: \"kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75\") pod \"neutron-058d-account-create-fzsds\" (UID: \"ce461b22-96d9-42ee-950d-b001054c5c60\") " pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.391978 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:03:58 crc kubenswrapper[4688]: I1001 16:03:58.440831 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-52dc-account-create-jqvjq"] Oct 01 16:03:58 crc kubenswrapper[4688]: W1001 16:03:58.452105 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59228674_5e5c_426f_9e2d_e82c5760b356.slice/crio-e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e WatchSource:0}: Error finding container e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e: Status 404 returned error can't find the container with id e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.583821 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d7e0-account-create-58kmq"] Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.978945 4688 generic.go:334] "Generic (PLEG): container finished" podID="59228674-5e5c-426f-9e2d-e82c5760b356" containerID="4c1182b103b63c55e821d4c81e0bc842973f8d71af838535e1aa1cd95606d7f5" exitCode=0 Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.979398 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-52dc-account-create-jqvjq" event={"ID":"59228674-5e5c-426f-9e2d-e82c5760b356","Type":"ContainerDied","Data":"4c1182b103b63c55e821d4c81e0bc842973f8d71af838535e1aa1cd95606d7f5"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.979432 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-52dc-account-create-jqvjq" event={"ID":"59228674-5e5c-426f-9e2d-e82c5760b356","Type":"ContainerStarted","Data":"e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.996868 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d7e0-account-create-58kmq" event={"ID":"f8c4d47a-f834-4952-a414-34fb8df2a1ef","Type":"ContainerStarted","Data":"1a7f60f52a6e0cf131857189106dac12916aaba97dd111722dbbb63576616a99"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:58.996915 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d7e0-account-create-58kmq" event={"ID":"f8c4d47a-f834-4952-a414-34fb8df2a1ef","Type":"ContainerStarted","Data":"2689b6aa21ab0137a3ed31114da92412ddaab4576dcbf00ece507c883345f165"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.008439 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"1f7bc70afabe0d67f368d34972f58be061e66b4dd2b96e12874aedd6452396e7"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.008482 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"8e9687f1be3ff39f52bb388b2f54ce561b40e4f5c53807b35d97d8ed03809d86"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.008495 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"897ee56172b4f9c41b4f9451e066427ba8a6147d836bbeeefc0fba58afe804dc"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.008504 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"9c285d8044485065da204a0abbe66424a381f220155df24353bfd8b96ecf386b"} Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.016855 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-d7e0-account-create-58kmq" podStartSLOduration=2.016836877 podStartE2EDuration="2.016836877s" podCreationTimestamp="2025-10-01 16:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:03:59.015592081 +0000 UTC m=+1028.366232043" watchObservedRunningTime="2025-10-01 16:03:59.016836877 +0000 UTC m=+1028.367476839" Oct 01 16:03:59 crc kubenswrapper[4688]: W1001 16:03:59.924401 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce461b22_96d9_42ee_950d_b001054c5c60.slice/crio-28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef WatchSource:0}: Error finding container 28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef: Status 404 returned error can't find the container with id 28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef Oct 01 16:03:59 crc kubenswrapper[4688]: I1001 16:03:59.930130 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-058d-account-create-fzsds"] Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.016546 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-058d-account-create-fzsds" event={"ID":"ce461b22-96d9-42ee-950d-b001054c5c60","Type":"ContainerStarted","Data":"28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef"} Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.034391 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"9195f2f3170430f92b3c72daa4d7d5001af3fe2154aaa1a00be3f960b378cf73"} Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.034442 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"d5ef1d73065cca99f2d89d49f1e2ec4a2382ec11d87a8dcfe19a778310324aab"} Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.034456 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0cd57226-8163-4319-ad42-5a17df19881a","Type":"ContainerStarted","Data":"ceee85e1ff2a99375f2fd0f57746e222b82d6607cf19faaf552a77aea78f357b"} Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.037210 4688 generic.go:334] "Generic (PLEG): container finished" podID="f8c4d47a-f834-4952-a414-34fb8df2a1ef" containerID="1a7f60f52a6e0cf131857189106dac12916aaba97dd111722dbbb63576616a99" exitCode=0 Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.037461 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d7e0-account-create-58kmq" event={"ID":"f8c4d47a-f834-4952-a414-34fb8df2a1ef","Type":"ContainerDied","Data":"1a7f60f52a6e0cf131857189106dac12916aaba97dd111722dbbb63576616a99"} Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.102099 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.654365894 podStartE2EDuration="48.102077142s" podCreationTimestamp="2025-10-01 16:03:12 +0000 UTC" firstStartedPulling="2025-10-01 16:03:46.410709528 +0000 UTC m=+1015.761349490" lastFinishedPulling="2025-10-01 16:03:57.858420776 +0000 UTC m=+1027.209060738" observedRunningTime="2025-10-01 16:04:00.088825781 +0000 UTC m=+1029.439465753" watchObservedRunningTime="2025-10-01 16:04:00.102077142 +0000 UTC m=+1029.452717344" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.364853 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.366870 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.373722 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.376096 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.461550 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.524587 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.524670 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.525149 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.525228 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.525296 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.525411 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzjvv\" (UniqueName: \"kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627175 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k95cr\" (UniqueName: \"kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr\") pod \"59228674-5e5c-426f-9e2d-e82c5760b356\" (UID: \"59228674-5e5c-426f-9e2d-e82c5760b356\") " Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627490 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627627 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzjvv\" (UniqueName: \"kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627670 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627741 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.627804 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.628806 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.629704 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.629707 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.629805 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.629954 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.649537 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr" (OuterVolumeSpecName: "kube-api-access-k95cr") pod "59228674-5e5c-426f-9e2d-e82c5760b356" (UID: "59228674-5e5c-426f-9e2d-e82c5760b356"). InnerVolumeSpecName "kube-api-access-k95cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.652664 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzjvv\" (UniqueName: \"kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv\") pod \"dnsmasq-dns-5c79d794d7-mwx5b\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.729190 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k95cr\" (UniqueName: \"kubernetes.io/projected/59228674-5e5c-426f-9e2d-e82c5760b356-kube-api-access-k95cr\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:00 crc kubenswrapper[4688]: I1001 16:04:00.787217 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.044262 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.046824 4688 generic.go:334] "Generic (PLEG): container finished" podID="ce461b22-96d9-42ee-950d-b001054c5c60" containerID="e280f20963ee60ed0f9b24bb156990c57adfec45b97ee9c49d49a285630e73fa" exitCode=0 Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.046886 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-058d-account-create-fzsds" event={"ID":"ce461b22-96d9-42ee-950d-b001054c5c60","Type":"ContainerDied","Data":"e280f20963ee60ed0f9b24bb156990c57adfec45b97ee9c49d49a285630e73fa"} Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.050681 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-52dc-account-create-jqvjq" event={"ID":"59228674-5e5c-426f-9e2d-e82c5760b356","Type":"ContainerDied","Data":"e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e"} Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.050741 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e55c5b21612ba60631bd383ac7df561e0d5f97a5e4a4a16d04c42f1d8b58294e" Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.050856 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-52dc-account-create-jqvjq" Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.313841 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.443975 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c28ql\" (UniqueName: \"kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql\") pod \"f8c4d47a-f834-4952-a414-34fb8df2a1ef\" (UID: \"f8c4d47a-f834-4952-a414-34fb8df2a1ef\") " Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.449088 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql" (OuterVolumeSpecName: "kube-api-access-c28ql") pod "f8c4d47a-f834-4952-a414-34fb8df2a1ef" (UID: "f8c4d47a-f834-4952-a414-34fb8df2a1ef"). InnerVolumeSpecName "kube-api-access-c28ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:01 crc kubenswrapper[4688]: I1001 16:04:01.546015 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c28ql\" (UniqueName: \"kubernetes.io/projected/f8c4d47a-f834-4952-a414-34fb8df2a1ef-kube-api-access-c28ql\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.059672 4688 generic.go:334] "Generic (PLEG): container finished" podID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerID="54bceb5de9646a9ccd4e389492e31ef4df5f3e106d0fd3b1d2ca846acc4ac415" exitCode=0 Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.059714 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" event={"ID":"cdfeab1a-74ca-413b-a15f-46f0f4085e80","Type":"ContainerDied","Data":"54bceb5de9646a9ccd4e389492e31ef4df5f3e106d0fd3b1d2ca846acc4ac415"} Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.060093 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" event={"ID":"cdfeab1a-74ca-413b-a15f-46f0f4085e80","Type":"ContainerStarted","Data":"a8a279565974837004203c435c6fd81e9eae952583b8663e4c215925c30d144e"} Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.062121 4688 generic.go:334] "Generic (PLEG): container finished" podID="e322d157-d4b5-4b2c-92a6-141481c2b358" containerID="de3e989d53a2da480bed2e6dae2272057e53c483fce3ba61e00724c73afff06b" exitCode=0 Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.062202 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bk86d" event={"ID":"e322d157-d4b5-4b2c-92a6-141481c2b358","Type":"ContainerDied","Data":"de3e989d53a2da480bed2e6dae2272057e53c483fce3ba61e00724c73afff06b"} Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.063914 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d7e0-account-create-58kmq" event={"ID":"f8c4d47a-f834-4952-a414-34fb8df2a1ef","Type":"ContainerDied","Data":"2689b6aa21ab0137a3ed31114da92412ddaab4576dcbf00ece507c883345f165"} Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.063967 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2689b6aa21ab0137a3ed31114da92412ddaab4576dcbf00ece507c883345f165" Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.063939 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d7e0-account-create-58kmq" Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.331696 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.366371 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwb75\" (UniqueName: \"kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75\") pod \"ce461b22-96d9-42ee-950d-b001054c5c60\" (UID: \"ce461b22-96d9-42ee-950d-b001054c5c60\") " Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.372576 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75" (OuterVolumeSpecName: "kube-api-access-rwb75") pod "ce461b22-96d9-42ee-950d-b001054c5c60" (UID: "ce461b22-96d9-42ee-950d-b001054c5c60"). InnerVolumeSpecName "kube-api-access-rwb75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:02 crc kubenswrapper[4688]: I1001 16:04:02.468626 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwb75\" (UniqueName: \"kubernetes.io/projected/ce461b22-96d9-42ee-950d-b001054c5c60-kube-api-access-rwb75\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.071757 4688 generic.go:334] "Generic (PLEG): container finished" podID="ad6973b9-8d39-441f-ae39-67b632e06bcf" containerID="46fcbc2ddeab50f23f49d9c431eb49e6459975bf3203d72e65912f21aa081c8a" exitCode=0 Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.071815 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l5md4" event={"ID":"ad6973b9-8d39-441f-ae39-67b632e06bcf","Type":"ContainerDied","Data":"46fcbc2ddeab50f23f49d9c431eb49e6459975bf3203d72e65912f21aa081c8a"} Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.074044 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-058d-account-create-fzsds" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.074047 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-058d-account-create-fzsds" event={"ID":"ce461b22-96d9-42ee-950d-b001054c5c60","Type":"ContainerDied","Data":"28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef"} Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.074215 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28b45c4e6df51aeeb6423c973d39b39996542c256e14812791a1d87a391a13ef" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.076096 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" event={"ID":"cdfeab1a-74ca-413b-a15f-46f0f4085e80","Type":"ContainerStarted","Data":"a0bf40881a1698b8b45a54295220e4c79ab817b55040474a68ff123ba9c178e4"} Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.076351 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.122444 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" podStartSLOduration=3.122422101 podStartE2EDuration="3.122422101s" podCreationTimestamp="2025-10-01 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:03.114414121 +0000 UTC m=+1032.465054093" watchObservedRunningTime="2025-10-01 16:04:03.122422101 +0000 UTC m=+1032.473062063" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.395901 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bk86d" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.582951 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle\") pod \"e322d157-d4b5-4b2c-92a6-141481c2b358\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.583110 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5wr4\" (UniqueName: \"kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4\") pod \"e322d157-d4b5-4b2c-92a6-141481c2b358\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.583729 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data\") pod \"e322d157-d4b5-4b2c-92a6-141481c2b358\" (UID: \"e322d157-d4b5-4b2c-92a6-141481c2b358\") " Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.587199 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4" (OuterVolumeSpecName: "kube-api-access-v5wr4") pod "e322d157-d4b5-4b2c-92a6-141481c2b358" (UID: "e322d157-d4b5-4b2c-92a6-141481c2b358"). InnerVolumeSpecName "kube-api-access-v5wr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.609626 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e322d157-d4b5-4b2c-92a6-141481c2b358" (UID: "e322d157-d4b5-4b2c-92a6-141481c2b358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.630587 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data" (OuterVolumeSpecName: "config-data") pod "e322d157-d4b5-4b2c-92a6-141481c2b358" (UID: "e322d157-d4b5-4b2c-92a6-141481c2b358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.686023 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.686230 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e322d157-d4b5-4b2c-92a6-141481c2b358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:03 crc kubenswrapper[4688]: I1001 16:04:03.686286 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5wr4\" (UniqueName: \"kubernetes.io/projected/e322d157-d4b5-4b2c-92a6-141481c2b358-kube-api-access-v5wr4\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.085590 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bk86d" event={"ID":"e322d157-d4b5-4b2c-92a6-141481c2b358","Type":"ContainerDied","Data":"cd94deb706654f4662eddbe2ffa353192586e9090dd2aa70c157756e3d1d50fb"} Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.085661 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd94deb706654f4662eddbe2ffa353192586e9090dd2aa70c157756e3d1d50fb" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.085633 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bk86d" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.374417 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.421436 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pg2qf"] Oct 01 16:04:04 crc kubenswrapper[4688]: E1001 16:04:04.421858 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e322d157-d4b5-4b2c-92a6-141481c2b358" containerName="keystone-db-sync" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.421876 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="e322d157-d4b5-4b2c-92a6-141481c2b358" containerName="keystone-db-sync" Oct 01 16:04:04 crc kubenswrapper[4688]: E1001 16:04:04.421891 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59228674-5e5c-426f-9e2d-e82c5760b356" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.421899 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="59228674-5e5c-426f-9e2d-e82c5760b356" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: E1001 16:04:04.421914 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c4d47a-f834-4952-a414-34fb8df2a1ef" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.421921 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c4d47a-f834-4952-a414-34fb8df2a1ef" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: E1001 16:04:04.421934 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce461b22-96d9-42ee-950d-b001054c5c60" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.421941 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce461b22-96d9-42ee-950d-b001054c5c60" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.422140 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="e322d157-d4b5-4b2c-92a6-141481c2b358" containerName="keystone-db-sync" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.422164 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce461b22-96d9-42ee-950d-b001054c5c60" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.422178 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="59228674-5e5c-426f-9e2d-e82c5760b356" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.422192 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c4d47a-f834-4952-a414-34fb8df2a1ef" containerName="mariadb-account-create" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.422788 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.433303 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.433596 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcpfl" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.440934 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.447860 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.517187 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.576876 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.593899 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pg2qf"] Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611134 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611364 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611422 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611446 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8g55\" (UniqueName: \"kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611473 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.611594 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.636276 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715417 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715466 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715489 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8g55\" (UniqueName: \"kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715508 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715542 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715581 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715600 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb27z\" (UniqueName: \"kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715625 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715646 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715663 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715722 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.715753 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.750110 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.751488 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.751925 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.771386 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.774080 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.798882 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8g55\" (UniqueName: \"kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55\") pod \"keystone-bootstrap-pg2qf\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.816990 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.817067 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.817105 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.817128 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.817162 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.817180 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb27z\" (UniqueName: \"kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.824640 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.825375 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.835079 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.836251 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.836413 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.888488 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l5md4" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.905762 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb27z\" (UniqueName: \"kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z\") pod \"dnsmasq-dns-5b868669f-pk2wq\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:04 crc kubenswrapper[4688]: I1001 16:04:04.931253 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.024281 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6br86\" (UniqueName: \"kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86\") pod \"ad6973b9-8d39-441f-ae39-67b632e06bcf\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.024799 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data\") pod \"ad6973b9-8d39-441f-ae39-67b632e06bcf\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.024930 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data\") pod \"ad6973b9-8d39-441f-ae39-67b632e06bcf\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.024989 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle\") pod \"ad6973b9-8d39-441f-ae39-67b632e06bcf\" (UID: \"ad6973b9-8d39-441f-ae39-67b632e06bcf\") " Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.040553 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86" (OuterVolumeSpecName: "kube-api-access-6br86") pod "ad6973b9-8d39-441f-ae39-67b632e06bcf" (UID: "ad6973b9-8d39-441f-ae39-67b632e06bcf"). InnerVolumeSpecName "kube-api-access-6br86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.057299 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ad6973b9-8d39-441f-ae39-67b632e06bcf" (UID: "ad6973b9-8d39-441f-ae39-67b632e06bcf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.075848 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:05 crc kubenswrapper[4688]: E1001 16:04:05.076271 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6973b9-8d39-441f-ae39-67b632e06bcf" containerName="glance-db-sync" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.076294 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6973b9-8d39-441f-ae39-67b632e06bcf" containerName="glance-db-sync" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.076547 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6973b9-8d39-441f-ae39-67b632e06bcf" containerName="glance-db-sync" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.077593 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.097727 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.115442 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.115702 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.115880 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.116018 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-brtxr" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.129837 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.135878 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.135927 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsz2n\" (UniqueName: \"kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.136022 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.136062 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.136191 4688 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.136203 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6br86\" (UniqueName: \"kubernetes.io/projected/ad6973b9-8d39-441f-ae39-67b632e06bcf-kube-api-access-6br86\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.147161 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.159794 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data" (OuterVolumeSpecName: "config-data") pod "ad6973b9-8d39-441f-ae39-67b632e06bcf" (UID: "ad6973b9-8d39-441f-ae39-67b632e06bcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.160696 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad6973b9-8d39-441f-ae39-67b632e06bcf" (UID: "ad6973b9-8d39-441f-ae39-67b632e06bcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.187925 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="dnsmasq-dns" containerID="cri-o://a0bf40881a1698b8b45a54295220e4c79ab817b55040474a68ff123ba9c178e4" gracePeriod=10 Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.188236 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l5md4" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.197020 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l5md4" event={"ID":"ad6973b9-8d39-441f-ae39-67b632e06bcf","Type":"ContainerDied","Data":"d285a82a34e0dad93f3561240365b102316c358cf0455a360f41098981c97420"} Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.197072 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d285a82a34e0dad93f3561240365b102316c358cf0455a360f41098981c97420" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240567 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240620 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsz2n\" (UniqueName: \"kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240669 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240781 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240877 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.240891 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6973b9-8d39-441f-ae39-67b632e06bcf-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.241861 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.242336 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.249853 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.253700 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.264735 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-bqw9p"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.265765 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.277099 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.277511 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.277777 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7jdv9" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.305470 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bqw9p"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.330351 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ntpwg"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.331478 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.335321 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-955dk" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.335563 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344452 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgsvx\" (UniqueName: \"kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344581 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344602 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344622 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344645 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.344662 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.347657 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsz2n\" (UniqueName: \"kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n\") pod \"horizon-7c9c47cbd9-tpv62\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.349688 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ntpwg"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.371899 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.407515 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.410621 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-bpllf"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.411917 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.445884 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.445953 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446009 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446027 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446057 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446089 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brgzp\" (UniqueName: \"kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446114 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446134 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446173 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gprb5\" (UniqueName: \"kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446209 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgsvx\" (UniqueName: \"kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446243 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446267 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446297 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446339 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.446743 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.451547 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.451733 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pldnx" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.454566 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bpllf"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.462509 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.462931 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.469067 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.469455 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.471917 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.479213 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgsvx\" (UniqueName: \"kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx\") pod \"cinder-db-sync-bqw9p\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.511425 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.524689 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547341 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547389 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547425 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547449 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547466 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs5tt\" (UniqueName: \"kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547487 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547505 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547546 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547569 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547587 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547617 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brgzp\" (UniqueName: \"kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547651 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gprb5\" (UniqueName: \"kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547685 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.547704 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.565811 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.566143 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.579401 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.580401 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.584827 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.597324 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.597717 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.614486 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gprb5\" (UniqueName: \"kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.648823 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.648860 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs5tt\" (UniqueName: \"kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.648882 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.648926 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.648989 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.649009 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.649805 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.655035 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brgzp\" (UniqueName: \"kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp\") pod \"placement-db-sync-ntpwg\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.655316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.656012 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.656842 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.657707 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.666700 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.667957 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.675054 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.677476 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle\") pod \"barbican-db-sync-bpllf\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.732990 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bpllf" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.737651 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.755663 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.756886 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs5tt\" (UniqueName: \"kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt\") pod \"dnsmasq-dns-cf78879c9-nrs5s\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.758073 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.758186 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.763959 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.764029 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.764065 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.764142 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhdt4\" (UniqueName: \"kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.764284 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.766013 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.766218 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.860725 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875237 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875311 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhdt4\" (UniqueName: \"kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875330 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875371 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875586 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875606 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875635 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875671 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875699 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875740 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875777 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.875807 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wbc\" (UniqueName: \"kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.879679 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.880825 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.882376 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.901686 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.925031 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhdt4\" (UniqueName: \"kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4\") pod \"horizon-5889fcf659-ptwxd\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995618 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995673 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995717 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995893 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995922 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.995974 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.996061 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wbc\" (UniqueName: \"kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:05 crc kubenswrapper[4688]: I1001 16:04:05.996302 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.003696 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.005970 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.024200 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.044362 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.044803 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.045449 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.090284 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wbc\" (UniqueName: \"kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc\") pod \"ceilometer-0\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.100397 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.186285 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.187758 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.227579 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.244969 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.248985 4688 generic.go:334] "Generic (PLEG): container finished" podID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerID="a0bf40881a1698b8b45a54295220e4c79ab817b55040474a68ff123ba9c178e4" exitCode=0 Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.249086 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" event={"ID":"cdfeab1a-74ca-413b-a15f-46f0f4085e80","Type":"ContainerDied","Data":"a0bf40881a1698b8b45a54295220e4c79ab817b55040474a68ff123ba9c178e4"} Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311384 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7fpq\" (UniqueName: \"kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311435 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311454 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311485 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311515 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.311618 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.342698 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.412685 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7fpq\" (UniqueName: \"kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.413001 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.413021 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.413051 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.413272 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.413394 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.414615 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.414687 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.414862 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.415387 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.420617 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.436938 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7fpq\" (UniqueName: \"kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq\") pod \"dnsmasq-dns-56df8fb6b7-r48br\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.580416 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.920227 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pg2qf"] Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.958388 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:06 crc kubenswrapper[4688]: W1001 16:04:06.974899 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7b044e0_4fc7_47c4_a7fd_deda5f608e8d.slice/crio-dd5f262b31c4cc247a56ddeed87f621f8617a8f5fe140424978918fcba76de96 WatchSource:0}: Error finding container dd5f262b31c4cc247a56ddeed87f621f8617a8f5fe140424978918fcba76de96: Status 404 returned error can't find the container with id dd5f262b31c4cc247a56ddeed87f621f8617a8f5fe140424978918fcba76de96 Oct 01 16:04:06 crc kubenswrapper[4688]: I1001 16:04:06.995075 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bpllf"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.007392 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bqw9p"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.037598 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ntpwg"] Oct 01 16:04:07 crc kubenswrapper[4688]: W1001 16:04:07.069339 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81b3c008_7340_436e_873b_aa0d41588c6c.slice/crio-607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d WatchSource:0}: Error finding container 607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d: Status 404 returned error can't find the container with id 607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d Oct 01 16:04:07 crc kubenswrapper[4688]: W1001 16:04:07.101665 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26e68b63_d053_412d_95e8_0e5f10bffc45.slice/crio-2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7 WatchSource:0}: Error finding container 2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7: Status 404 returned error can't find the container with id 2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7 Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.177726 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.237466 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: E1001 16:04:07.237897 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="init" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.237915 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="init" Oct 01 16:04:07 crc kubenswrapper[4688]: E1001 16:04:07.237954 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="dnsmasq-dns" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.237961 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="dnsmasq-dns" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.238158 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" containerName="dnsmasq-dns" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.239007 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.244039 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.244163 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.244203 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pbr6z" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.250875 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.289188 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9c47cbd9-tpv62" event={"ID":"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d","Type":"ContainerStarted","Data":"dd5f262b31c4cc247a56ddeed87f621f8617a8f5fe140424978918fcba76de96"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.290277 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ntpwg" event={"ID":"26e68b63-d053-412d-95e8-0e5f10bffc45","Type":"ContainerStarted","Data":"2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.302673 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pg2qf" event={"ID":"ac60a38f-7df7-4586-a18a-d48fdd1e12b8","Type":"ContainerStarted","Data":"09732cecad4f143ebc132270d6296fe9fe9f2747f5db93fa29ac48211ab5dcfc"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.305266 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bpllf" event={"ID":"8ff202f4-494c-49d4-a228-2e85d6659bb9","Type":"ContainerStarted","Data":"6d2f41598d0ec8132a19ce85408af5e6624a4071db79cc403cbfde8d9e9d9358"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.318606 4688 generic.go:334] "Generic (PLEG): container finished" podID="6fe9fd2d-1776-4994-9fd8-b90060cd19f7" containerID="33349de2ba8915ab05e34b50dc1cead8e0187dae0dd4138f5cb2c012daac3f7c" exitCode=0 Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.318716 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" event={"ID":"6fe9fd2d-1776-4994-9fd8-b90060cd19f7","Type":"ContainerDied","Data":"33349de2ba8915ab05e34b50dc1cead8e0187dae0dd4138f5cb2c012daac3f7c"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.318748 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" event={"ID":"6fe9fd2d-1776-4994-9fd8-b90060cd19f7","Type":"ContainerStarted","Data":"74e657c993e84dba0ece4d6a7e18b076b5f3c85e402413b0a3fbb0db50a3232f"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.325629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" event={"ID":"cdfeab1a-74ca-413b-a15f-46f0f4085e80","Type":"ContainerDied","Data":"a8a279565974837004203c435c6fd81e9eae952583b8663e4c215925c30d144e"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.325669 4688 scope.go:117] "RemoveContainer" containerID="a0bf40881a1698b8b45a54295220e4c79ab817b55040474a68ff123ba9c178e4" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.325786 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-mwx5b" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.327572 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqw9p" event={"ID":"81b3c008-7340-436e-873b-aa0d41588c6c","Type":"ContainerStarted","Data":"607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d"} Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338639 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzjvv\" (UniqueName: \"kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338727 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338789 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338817 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338855 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.338874 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb\") pod \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\" (UID: \"cdfeab1a-74ca-413b-a15f-46f0f4085e80\") " Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339162 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfk9k\" (UniqueName: \"kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339201 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339221 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339262 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339286 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339357 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.339421 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.355035 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv" (OuterVolumeSpecName: "kube-api-access-bzjvv") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "kube-api-access-bzjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.416935 4688 scope.go:117] "RemoveContainer" containerID="54bceb5de9646a9ccd4e389492e31ef4df5f3e106d0fd3b1d2ca846acc4ac415" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443021 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443135 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443207 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443266 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfk9k\" (UniqueName: \"kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443308 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443330 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.443413 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzjvv\" (UniqueName: \"kubernetes.io/projected/cdfeab1a-74ca-413b-a15f-46f0f4085e80-kube-api-access-bzjvv\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.445230 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.447293 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.451869 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.468473 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.468558 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.475100 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.483340 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.496126 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfk9k\" (UniqueName: \"kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.512313 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.517187 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.525212 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.525408 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.539333 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.545958 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550045 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550138 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550214 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550261 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550336 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550406 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.550464 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq8g2\" (UniqueName: \"kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.583288 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664348 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664725 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq8g2\" (UniqueName: \"kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664776 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664804 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664870 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664899 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.664961 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.665387 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.665669 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.666101 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.715439 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq8g2\" (UniqueName: \"kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.731056 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.745558 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: E1001 16:04:07.745980 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="227029d5-aefb-4213-8d5a-bd579eeac5f9" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.760623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: W1001 16:04:07.792140 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e1915c3_1db9_4c2f_8011_df62c28827af.slice/crio-58523a79bc353c3b058ae84a9c282ad2bf651b68a72d02eb01f27c86f4ac00ae WatchSource:0}: Error finding container 58523a79bc353c3b058ae84a9c282ad2bf651b68a72d02eb01f27c86f4ac00ae: Status 404 returned error can't find the container with id 58523a79bc353c3b058ae84a9c282ad2bf651b68a72d02eb01f27c86f4ac00ae Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.804639 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.818638 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.830596 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.850245 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.851868 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.878211 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.878271 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.878303 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.878400 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8k5\" (UniqueName: \"kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.878447 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.906096 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.922236 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: E1001 16:04:07.922903 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="414cd993-de99-43f4-8ebd-5b17b7961a85" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.966645 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.980948 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.980988 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.981017 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.981084 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8k5\" (UniqueName: \"kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.981128 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.982122 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.982673 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.983466 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:07 crc kubenswrapper[4688]: I1001 16:04:07.990454 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.007154 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.011952 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8k5\" (UniqueName: \"kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5\") pod \"horizon-58744c64d9-c94d2\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.030721 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.076863 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.084928 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.084958 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.084970 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.088154 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.153365 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config" (OuterVolumeSpecName: "config") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.168094 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cdfeab1a-74ca-413b-a15f-46f0f4085e80" (UID: "cdfeab1a-74ca-413b-a15f-46f0f4085e80"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.186592 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.186636 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdfeab1a-74ca-413b-a15f-46f0f4085e80-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.296741 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.299470 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.305656 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-mwx5b"] Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.310115 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.347765 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pg2qf" event={"ID":"ac60a38f-7df7-4586-a18a-d48fdd1e12b8","Type":"ContainerStarted","Data":"87e25ddf63b34ce84c3833804485adea9b8b919ad25c39fd02707d4797968acb"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.359735 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" event={"ID":"7e1915c3-1db9-4c2f-8011-df62c28827af","Type":"ContainerStarted","Data":"58523a79bc353c3b058ae84a9c282ad2bf651b68a72d02eb01f27c86f4ac00ae"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.361719 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerStarted","Data":"414ed783e209ce593cad2d7110a77218656be1b3c6e8f544432424a7fb2443dc"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.369102 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerStarted","Data":"bbdd091dad51544a67eb44e3ba2a3aba08ab1fcdd81e9a58edf44b4ab68f597a"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.374916 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" event={"ID":"6fe9fd2d-1776-4994-9fd8-b90060cd19f7","Type":"ContainerDied","Data":"74e657c993e84dba0ece4d6a7e18b076b5f3c85e402413b0a3fbb0db50a3232f"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.374981 4688 scope.go:117] "RemoveContainer" containerID="33349de2ba8915ab05e34b50dc1cead8e0187dae0dd4138f5cb2c012daac3f7c" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.375217 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-pk2wq" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.392723 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.392802 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.392828 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb27z\" (UniqueName: \"kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.392872 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.392997 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.393047 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config\") pod \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\" (UID: \"6fe9fd2d-1776-4994-9fd8-b90060cd19f7\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.417728 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.418315 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" event={"ID":"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb","Type":"ContainerStarted","Data":"5e7c4aa084730a5a4f95d070313c22e7a4b0e3b972964e089d041d9d08096c00"} Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.418362 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.425472 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pg2qf" podStartSLOduration=4.425453022 podStartE2EDuration="4.425453022s" podCreationTimestamp="2025-10-01 16:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:08.369450384 +0000 UTC m=+1037.720090346" watchObservedRunningTime="2025-10-01 16:04:08.425453022 +0000 UTC m=+1037.776092984" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.428254 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-flx64"] Oct 01 16:04:08 crc kubenswrapper[4688]: E1001 16:04:08.428615 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe9fd2d-1776-4994-9fd8-b90060cd19f7" containerName="init" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.428633 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe9fd2d-1776-4994-9fd8-b90060cd19f7" containerName="init" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.428813 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe9fd2d-1776-4994-9fd8-b90060cd19f7" containerName="init" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.429384 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.431335 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z" (OuterVolumeSpecName: "kube-api-access-tb27z") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "kube-api-access-tb27z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.433713 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qzsnx" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.433956 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.449815 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.456025 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-flx64"] Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.459035 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config" (OuterVolumeSpecName: "config") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.467252 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496058 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw95n\" (UniqueName: \"kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496113 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496232 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496346 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496365 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.496384 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb27z\" (UniqueName: \"kubernetes.io/projected/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-kube-api-access-tb27z\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.501593 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.512734 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.520336 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.533329 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.543026 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6fe9fd2d-1776-4994-9fd8-b90060cd19f7" (UID: "6fe9fd2d-1776-4994-9fd8-b90060cd19f7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.598929 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfk9k\" (UniqueName: \"kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.598978 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599006 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599045 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599088 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599110 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599179 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599203 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599247 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599265 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599292 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599313 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599359 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq8g2\" (UniqueName: \"kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2\") pod \"414cd993-de99-43f4-8ebd-5b17b7961a85\" (UID: \"414cd993-de99-43f4-8ebd-5b17b7961a85\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599375 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run\") pod \"227029d5-aefb-4213-8d5a-bd579eeac5f9\" (UID: \"227029d5-aefb-4213-8d5a-bd579eeac5f9\") " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599652 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599729 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw95n\" (UniqueName: \"kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599768 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599809 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599820 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.599832 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe9fd2d-1776-4994-9fd8-b90060cd19f7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.604457 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs" (OuterVolumeSpecName: "logs") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.604655 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.611001 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs" (OuterVolumeSpecName: "logs") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.611250 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.618632 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts" (OuterVolumeSpecName: "scripts") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.619560 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k" (OuterVolumeSpecName: "kube-api-access-cfk9k") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "kube-api-access-cfk9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.619678 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.619732 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.623178 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.624831 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.624896 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data" (OuterVolumeSpecName: "config-data") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.624938 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts" (OuterVolumeSpecName: "scripts") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.625200 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2" (OuterVolumeSpecName: "kube-api-access-jq8g2") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "kube-api-access-jq8g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.625444 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "414cd993-de99-43f4-8ebd-5b17b7961a85" (UID: "414cd993-de99-43f4-8ebd-5b17b7961a85"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.625514 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data" (OuterVolumeSpecName: "config-data") pod "227029d5-aefb-4213-8d5a-bd579eeac5f9" (UID: "227029d5-aefb-4213-8d5a-bd579eeac5f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.637747 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.652217 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw95n\" (UniqueName: \"kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n\") pod \"neutron-db-sync-flx64\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.701583 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.701618 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.702953 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705250 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705278 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/414cd993-de99-43f4-8ebd-5b17b7961a85-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705287 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/227029d5-aefb-4213-8d5a-bd579eeac5f9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705296 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq8g2\" (UniqueName: \"kubernetes.io/projected/414cd993-de99-43f4-8ebd-5b17b7961a85-kube-api-access-jq8g2\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705305 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfk9k\" (UniqueName: \"kubernetes.io/projected/227029d5-aefb-4213-8d5a-bd579eeac5f9-kube-api-access-cfk9k\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705325 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705334 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705342 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705351 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705361 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/414cd993-de99-43f4-8ebd-5b17b7961a85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.705369 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/227029d5-aefb-4213-8d5a-bd579eeac5f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.751309 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.765883 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.821248 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.821287 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.866228 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.873187 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.913733 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-pk2wq"] Oct 01 16:04:08 crc kubenswrapper[4688]: I1001 16:04:08.961734 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:08 crc kubenswrapper[4688]: W1001 16:04:08.990265 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bf07c8b_8b89_473e_a1dc_f008cefe2be4.slice/crio-1d5eac5ee2ed41d8a6d447f20d471471b21765350a9dd9e21ac17f4f7b54c1e3 WatchSource:0}: Error finding container 1d5eac5ee2ed41d8a6d447f20d471471b21765350a9dd9e21ac17f4f7b54c1e3: Status 404 returned error can't find the container with id 1d5eac5ee2ed41d8a6d447f20d471471b21765350a9dd9e21ac17f4f7b54c1e3 Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.407283 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe9fd2d-1776-4994-9fd8-b90060cd19f7" path="/var/lib/kubelet/pods/6fe9fd2d-1776-4994-9fd8-b90060cd19f7/volumes" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.408239 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdfeab1a-74ca-413b-a15f-46f0f4085e80" path="/var/lib/kubelet/pods/cdfeab1a-74ca-413b-a15f-46f0f4085e80/volumes" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.452614 4688 generic.go:334] "Generic (PLEG): container finished" podID="89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" containerID="0fad19a1dae70f7d261ee01d56bd6ad7791c76cb5aba2a9ed560d7fca40b340c" exitCode=0 Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.452682 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" event={"ID":"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb","Type":"ContainerDied","Data":"0fad19a1dae70f7d261ee01d56bd6ad7791c76cb5aba2a9ed560d7fca40b340c"} Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.457822 4688 generic.go:334] "Generic (PLEG): container finished" podID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerID="d68f0020f5402bff3d6afb5d51c93e9f48ae69749b685812a5dea4b67412f4b0" exitCode=0 Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.457898 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" event={"ID":"7e1915c3-1db9-4c2f-8011-df62c28827af","Type":"ContainerDied","Data":"d68f0020f5402bff3d6afb5d51c93e9f48ae69749b685812a5dea4b67412f4b0"} Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.461592 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.461592 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58744c64d9-c94d2" event={"ID":"1bf07c8b-8b89-473e-a1dc-f008cefe2be4","Type":"ContainerStarted","Data":"1d5eac5ee2ed41d8a6d447f20d471471b21765350a9dd9e21ac17f4f7b54c1e3"} Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.461646 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.527486 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-flx64"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.600847 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.617372 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.658592 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.660143 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.672680 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.672917 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.673120 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pbr6z" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.712603 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.759261 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773332 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773402 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773489 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773565 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773597 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6cxt\" (UniqueName: \"kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773656 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.773674 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.788935 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.808449 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.809896 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.820739 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.827667 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876431 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876494 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876544 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6cxt\" (UniqueName: \"kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876577 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876596 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876632 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876656 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.876969 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.878766 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.879656 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.892053 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.895423 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.899066 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.903301 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6cxt\" (UniqueName: \"kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.951766 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.977872 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx25q\" (UniqueName: \"kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.977946 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.978001 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.978034 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.978070 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.978111 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:09 crc kubenswrapper[4688]: I1001 16:04:09.978134 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.031750 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079659 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079738 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079782 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079828 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079853 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079920 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx25q\" (UniqueName: \"kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.079959 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.080100 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.081897 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.082358 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.087755 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.090680 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.096374 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.100628 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx25q\" (UniqueName: \"kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.161603 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.203663 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290104 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290174 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290234 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs5tt\" (UniqueName: \"kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290264 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290326 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.290354 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0\") pod \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\" (UID: \"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb\") " Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.430813 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt" (OuterVolumeSpecName: "kube-api-access-fs5tt") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "kube-api-access-fs5tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.448477 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.490427 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-flx64" event={"ID":"4a989728-c717-4967-a372-93152bd2f064","Type":"ContainerStarted","Data":"970c8c2d9e9c586d88e639a9f8ca2f73cb1fa621ef664b649c343a5f3766000d"} Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.499970 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs5tt\" (UniqueName: \"kubernetes.io/projected/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-kube-api-access-fs5tt\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.515789 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" event={"ID":"89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb","Type":"ContainerDied","Data":"5e7c4aa084730a5a4f95d070313c22e7a4b0e3b972964e089d041d9d08096c00"} Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.515859 4688 scope.go:117] "RemoveContainer" containerID="0fad19a1dae70f7d261ee01d56bd6ad7791c76cb5aba2a9ed560d7fca40b340c" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.516029 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-nrs5s" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.748775 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.752222 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.785310 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.803996 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.804298 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.837841 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.871216 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: W1001 16:04:10.872301 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b37bdf8_ac75_459c_92f9_94e6b9b2d05c.slice/crio-dcea04d7232475de2626de734db0fd95c235304a8b4978b3eb6a9d2f219854b6 WatchSource:0}: Error finding container dcea04d7232475de2626de734db0fd95c235304a8b4978b3eb6a9d2f219854b6: Status 404 returned error can't find the container with id dcea04d7232475de2626de734db0fd95c235304a8b4978b3eb6a9d2f219854b6 Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.875777 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config" (OuterVolumeSpecName: "config") pod "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" (UID: "89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.906105 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.906134 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:10 crc kubenswrapper[4688]: I1001 16:04:10.906146 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.217319 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.221379 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-nrs5s"] Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.303358 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:11 crc kubenswrapper[4688]: W1001 16:04:11.318449 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b38e310_fa47_4eb6_8256_475f992c1efa.slice/crio-3b83075adfd37c288c0943c6b140eb9e5a89c4192e5f3a9b56395731bb688d71 WatchSource:0}: Error finding container 3b83075adfd37c288c0943c6b140eb9e5a89c4192e5f3a9b56395731bb688d71: Status 404 returned error can't find the container with id 3b83075adfd37c288c0943c6b140eb9e5a89c4192e5f3a9b56395731bb688d71 Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.395016 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="227029d5-aefb-4213-8d5a-bd579eeac5f9" path="/var/lib/kubelet/pods/227029d5-aefb-4213-8d5a-bd579eeac5f9/volumes" Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.395468 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="414cd993-de99-43f4-8ebd-5b17b7961a85" path="/var/lib/kubelet/pods/414cd993-de99-43f4-8ebd-5b17b7961a85/volumes" Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.395817 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" path="/var/lib/kubelet/pods/89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb/volumes" Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.598783 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerStarted","Data":"dcea04d7232475de2626de734db0fd95c235304a8b4978b3eb6a9d2f219854b6"} Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.628562 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-flx64" event={"ID":"4a989728-c717-4967-a372-93152bd2f064","Type":"ContainerStarted","Data":"8c779883d99658a75782ccb222ae4e75028da3290c76740c0708a162626cab76"} Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.632828 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerStarted","Data":"3b83075adfd37c288c0943c6b140eb9e5a89c4192e5f3a9b56395731bb688d71"} Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.660010 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-flx64" podStartSLOduration=3.659984771 podStartE2EDuration="3.659984771s" podCreationTimestamp="2025-10-01 16:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:11.658448728 +0000 UTC m=+1041.009088690" watchObservedRunningTime="2025-10-01 16:04:11.659984771 +0000 UTC m=+1041.010624733" Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.669797 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" event={"ID":"7e1915c3-1db9-4c2f-8011-df62c28827af","Type":"ContainerStarted","Data":"1f14ba0ef31a4de3533a673ab57ef7043aeb05057ada91913157bcbf4345a8a1"} Oct 01 16:04:11 crc kubenswrapper[4688]: I1001 16:04:11.670335 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:12 crc kubenswrapper[4688]: I1001 16:04:12.689986 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerStarted","Data":"3162bf0e25e7d61e27cf4f11caf9dbb1fe4462353ddad47ee4284f48988db10c"} Oct 01 16:04:12 crc kubenswrapper[4688]: I1001 16:04:12.692700 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerStarted","Data":"fb0f2877979d065e9427d7f77859cc896504e685b8d7529195c0334076f38602"} Oct 01 16:04:13 crc kubenswrapper[4688]: I1001 16:04:13.706616 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerStarted","Data":"e371a796da46d1f12192e2cb38d4b16d78038d5ff715684faa42c5c40ba709ca"} Oct 01 16:04:13 crc kubenswrapper[4688]: I1001 16:04:13.761950 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.761928748 podStartE2EDuration="4.761928748s" podCreationTimestamp="2025-10-01 16:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:13.756591049 +0000 UTC m=+1043.107231001" watchObservedRunningTime="2025-10-01 16:04:13.761928748 +0000 UTC m=+1043.112568710" Oct 01 16:04:13 crc kubenswrapper[4688]: I1001 16:04:13.763949 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" podStartSLOduration=7.763940954 podStartE2EDuration="7.763940954s" podCreationTimestamp="2025-10-01 16:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:11.701029863 +0000 UTC m=+1041.051669845" watchObservedRunningTime="2025-10-01 16:04:13.763940954 +0000 UTC m=+1043.114580916" Oct 01 16:04:14 crc kubenswrapper[4688]: I1001 16:04:14.739620 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerStarted","Data":"2f3cebbe63461991cc75c9da6c217efcdc60e4b2522df53c8a704fe1dd1017e1"} Oct 01 16:04:14 crc kubenswrapper[4688]: I1001 16:04:14.741935 4688 generic.go:334] "Generic (PLEG): container finished" podID="ac60a38f-7df7-4586-a18a-d48fdd1e12b8" containerID="87e25ddf63b34ce84c3833804485adea9b8b919ad25c39fd02707d4797968acb" exitCode=0 Oct 01 16:04:14 crc kubenswrapper[4688]: I1001 16:04:14.742915 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pg2qf" event={"ID":"ac60a38f-7df7-4586-a18a-d48fdd1e12b8","Type":"ContainerDied","Data":"87e25ddf63b34ce84c3833804485adea9b8b919ad25c39fd02707d4797968acb"} Oct 01 16:04:14 crc kubenswrapper[4688]: I1001 16:04:14.773685 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.77367086 podStartE2EDuration="5.77367086s" podCreationTimestamp="2025-10-01 16:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:14.768136806 +0000 UTC m=+1044.118776788" watchObservedRunningTime="2025-10-01 16:04:14.77367086 +0000 UTC m=+1044.124310822" Oct 01 16:04:16 crc kubenswrapper[4688]: I1001 16:04:16.582746 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:16 crc kubenswrapper[4688]: I1001 16:04:16.704057 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:04:16 crc kubenswrapper[4688]: I1001 16:04:16.704312 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" containerID="cri-o://e7b9aa1bdf9f1fb38248175acc41dd1be51e9764e75233db32aa665d91636bbc" gracePeriod=10 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.312069 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.312657 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-log" containerID="cri-o://fb0f2877979d065e9427d7f77859cc896504e685b8d7529195c0334076f38602" gracePeriod=30 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.312832 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-httpd" containerID="cri-o://2f3cebbe63461991cc75c9da6c217efcdc60e4b2522df53c8a704fe1dd1017e1" gracePeriod=30 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.448161 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.448437 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-log" containerID="cri-o://3162bf0e25e7d61e27cf4f11caf9dbb1fe4462353ddad47ee4284f48988db10c" gracePeriod=30 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.448981 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-httpd" containerID="cri-o://e371a796da46d1f12192e2cb38d4b16d78038d5ff715684faa42c5c40ba709ca" gracePeriod=30 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.783795 4688 generic.go:334] "Generic (PLEG): container finished" podID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerID="e371a796da46d1f12192e2cb38d4b16d78038d5ff715684faa42c5c40ba709ca" exitCode=0 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.783829 4688 generic.go:334] "Generic (PLEG): container finished" podID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerID="3162bf0e25e7d61e27cf4f11caf9dbb1fe4462353ddad47ee4284f48988db10c" exitCode=143 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.783862 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerDied","Data":"e371a796da46d1f12192e2cb38d4b16d78038d5ff715684faa42c5c40ba709ca"} Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.783885 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerDied","Data":"3162bf0e25e7d61e27cf4f11caf9dbb1fe4462353ddad47ee4284f48988db10c"} Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.785799 4688 generic.go:334] "Generic (PLEG): container finished" podID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerID="e7b9aa1bdf9f1fb38248175acc41dd1be51e9764e75233db32aa665d91636bbc" exitCode=0 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.785837 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" event={"ID":"62e6228f-ade9-4baf-a01f-8d5e19e5836c","Type":"ContainerDied","Data":"e7b9aa1bdf9f1fb38248175acc41dd1be51e9764e75233db32aa665d91636bbc"} Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.788550 4688 generic.go:334] "Generic (PLEG): container finished" podID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerID="2f3cebbe63461991cc75c9da6c217efcdc60e4b2522df53c8a704fe1dd1017e1" exitCode=0 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.788568 4688 generic.go:334] "Generic (PLEG): container finished" podID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerID="fb0f2877979d065e9427d7f77859cc896504e685b8d7529195c0334076f38602" exitCode=143 Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.788581 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerDied","Data":"2f3cebbe63461991cc75c9da6c217efcdc60e4b2522df53c8a704fe1dd1017e1"} Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.788596 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerDied","Data":"fb0f2877979d065e9427d7f77859cc896504e685b8d7529195c0334076f38602"} Oct 01 16:04:17 crc kubenswrapper[4688]: I1001 16:04:17.919762 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: connect: connection refused" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.022801 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.060932 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:04:18 crc kubenswrapper[4688]: E1001 16:04:18.061314 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" containerName="init" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.061336 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" containerName="init" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.061514 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="89746ed9-2bdc-48e7-aaa8-3a943aaf3dfb" containerName="init" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.062551 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.088918 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.103005 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124355 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124508 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124610 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124707 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124798 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124864 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.124946 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwnr\" (UniqueName: \"kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.185121 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235736 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235806 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235842 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235877 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235906 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235927 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.235956 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwnr\" (UniqueName: \"kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.236645 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.237824 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.242114 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-555656bf5b-xlntq"] Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.244179 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.245683 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.246441 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.246769 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.264665 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.267927 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.268084 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwnr\" (UniqueName: \"kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr\") pod \"horizon-654b94b7dd-g2k9q\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.305430 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-555656bf5b-xlntq"] Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337511 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337668 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337715 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8g55\" (UniqueName: \"kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337737 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337756 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337780 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts\") pod \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\" (UID: \"ac60a38f-7df7-4586-a18a-d48fdd1e12b8\") " Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.337998 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-secret-key\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338048 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-scripts\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338084 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-tls-certs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338171 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-config-data\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338217 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmb4\" (UniqueName: \"kubernetes.io/projected/37923e92-dbcc-41a1-8d2f-89d8de59959e-kube-api-access-vdmb4\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338242 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37923e92-dbcc-41a1-8d2f-89d8de59959e-logs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.338257 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-combined-ca-bundle\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.352141 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.352318 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.361709 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55" (OuterVolumeSpecName: "kube-api-access-r8g55") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "kube-api-access-r8g55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.371558 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts" (OuterVolumeSpecName: "scripts") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.379706 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.395180 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data" (OuterVolumeSpecName: "config-data") pod "ac60a38f-7df7-4586-a18a-d48fdd1e12b8" (UID: "ac60a38f-7df7-4586-a18a-d48fdd1e12b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.439895 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-secret-key\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440248 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-scripts\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440289 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-tls-certs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440379 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-config-data\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440423 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmb4\" (UniqueName: \"kubernetes.io/projected/37923e92-dbcc-41a1-8d2f-89d8de59959e-kube-api-access-vdmb4\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440456 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37923e92-dbcc-41a1-8d2f-89d8de59959e-logs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440475 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-combined-ca-bundle\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440567 4688 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440578 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8g55\" (UniqueName: \"kubernetes.io/projected/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-kube-api-access-r8g55\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440589 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440597 4688 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440605 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.440613 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac60a38f-7df7-4586-a18a-d48fdd1e12b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.443393 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37923e92-dbcc-41a1-8d2f-89d8de59959e-logs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.446023 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-scripts\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.446212 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37923e92-dbcc-41a1-8d2f-89d8de59959e-config-data\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.446554 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-secret-key\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.446670 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-combined-ca-bundle\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.456559 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/37923e92-dbcc-41a1-8d2f-89d8de59959e-horizon-tls-certs\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.470043 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmb4\" (UniqueName: \"kubernetes.io/projected/37923e92-dbcc-41a1-8d2f-89d8de59959e-kube-api-access-vdmb4\") pod \"horizon-555656bf5b-xlntq\" (UID: \"37923e92-dbcc-41a1-8d2f-89d8de59959e\") " pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.537060 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.586878 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.800133 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pg2qf" event={"ID":"ac60a38f-7df7-4586-a18a-d48fdd1e12b8","Type":"ContainerDied","Data":"09732cecad4f143ebc132270d6296fe9fe9f2747f5db93fa29ac48211ab5dcfc"} Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.800171 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09732cecad4f143ebc132270d6296fe9fe9f2747f5db93fa29ac48211ab5dcfc" Oct 01 16:04:18 crc kubenswrapper[4688]: I1001 16:04:18.800196 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pg2qf" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.437588 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pg2qf"] Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.439286 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pg2qf"] Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.537584 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-82bf8"] Oct 01 16:04:19 crc kubenswrapper[4688]: E1001 16:04:19.537982 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac60a38f-7df7-4586-a18a-d48fdd1e12b8" containerName="keystone-bootstrap" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.538003 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac60a38f-7df7-4586-a18a-d48fdd1e12b8" containerName="keystone-bootstrap" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.538194 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac60a38f-7df7-4586-a18a-d48fdd1e12b8" containerName="keystone-bootstrap" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.538886 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.541115 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.541735 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.542656 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.549296 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-82bf8"] Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.554011 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcpfl" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563100 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p69v4\" (UniqueName: \"kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563140 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563196 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563242 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563262 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.563379 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664727 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664786 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664806 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664833 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664902 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p69v4\" (UniqueName: \"kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.664922 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.670494 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.672667 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.672764 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.678799 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.687031 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p69v4\" (UniqueName: \"kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.687128 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle\") pod \"keystone-bootstrap-82bf8\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:19 crc kubenswrapper[4688]: I1001 16:04:19.859047 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:21 crc kubenswrapper[4688]: I1001 16:04:21.401732 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac60a38f-7df7-4586-a18a-d48fdd1e12b8" path="/var/lib/kubelet/pods/ac60a38f-7df7-4586-a18a-d48fdd1e12b8/volumes" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.094064 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.097268 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204137 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204210 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204276 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc\") pod \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204303 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204397 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204486 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx25q\" (UniqueName: \"kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204591 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205011 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs" (OuterVolumeSpecName: "logs") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205138 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.204620 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb\") pod \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205735 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config\") pod \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205775 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb\") pod \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205806 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgt8n\" (UniqueName: \"kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n\") pod \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\" (UID: \"62e6228f-ade9-4baf-a01f-8d5e19e5836c\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.205840 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle\") pod \"8b38e310-fa47-4eb6-8256-475f992c1efa\" (UID: \"8b38e310-fa47-4eb6-8256-475f992c1efa\") " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.207484 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.207622 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b38e310-fa47-4eb6-8256-475f992c1efa-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.214320 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts" (OuterVolumeSpecName: "scripts") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.227802 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.228955 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q" (OuterVolumeSpecName: "kube-api-access-xx25q") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "kube-api-access-xx25q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.244350 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.257978 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n" (OuterVolumeSpecName: "kube-api-access-pgt8n") pod "62e6228f-ade9-4baf-a01f-8d5e19e5836c" (UID: "62e6228f-ade9-4baf-a01f-8d5e19e5836c"). InnerVolumeSpecName "kube-api-access-pgt8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.276853 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data" (OuterVolumeSpecName: "config-data") pod "8b38e310-fa47-4eb6-8256-475f992c1efa" (UID: "8b38e310-fa47-4eb6-8256-475f992c1efa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.280736 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "62e6228f-ade9-4baf-a01f-8d5e19e5836c" (UID: "62e6228f-ade9-4baf-a01f-8d5e19e5836c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.285915 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "62e6228f-ade9-4baf-a01f-8d5e19e5836c" (UID: "62e6228f-ade9-4baf-a01f-8d5e19e5836c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.294688 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "62e6228f-ade9-4baf-a01f-8d5e19e5836c" (UID: "62e6228f-ade9-4baf-a01f-8d5e19e5836c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.306981 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config" (OuterVolumeSpecName: "config") pod "62e6228f-ade9-4baf-a01f-8d5e19e5836c" (UID: "62e6228f-ade9-4baf-a01f-8d5e19e5836c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309502 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx25q\" (UniqueName: \"kubernetes.io/projected/8b38e310-fa47-4eb6-8256-475f992c1efa-kube-api-access-xx25q\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309548 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309559 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309569 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309577 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgt8n\" (UniqueName: \"kubernetes.io/projected/62e6228f-ade9-4baf-a01f-8d5e19e5836c-kube-api-access-pgt8n\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309586 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309606 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309617 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309625 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62e6228f-ade9-4baf-a01f-8d5e19e5836c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.309633 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b38e310-fa47-4eb6-8256-475f992c1efa-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.333068 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.412818 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.868258 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8b38e310-fa47-4eb6-8256-475f992c1efa","Type":"ContainerDied","Data":"3b83075adfd37c288c0943c6b140eb9e5a89c4192e5f3a9b56395731bb688d71"} Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.868289 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.868329 4688 scope.go:117] "RemoveContainer" containerID="2f3cebbe63461991cc75c9da6c217efcdc60e4b2522df53c8a704fe1dd1017e1" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.871154 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" event={"ID":"62e6228f-ade9-4baf-a01f-8d5e19e5836c","Type":"ContainerDied","Data":"f3e8016c1d889b9dbf2e9590590ffed501c95788f88aea04a1dbd3d3a8981fcf"} Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.871260 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.896951 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.903799 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-77s2p"] Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.917430 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.933629 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944266 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:25 crc kubenswrapper[4688]: E1001 16:04:25.944641 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="init" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944658 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="init" Oct 01 16:04:25 crc kubenswrapper[4688]: E1001 16:04:25.944670 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-httpd" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944676 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-httpd" Oct 01 16:04:25 crc kubenswrapper[4688]: E1001 16:04:25.944690 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944697 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" Oct 01 16:04:25 crc kubenswrapper[4688]: E1001 16:04:25.944708 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-log" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944715 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-log" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944881 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-httpd" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944898 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" containerName="glance-log" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.944912 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.946171 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.948713 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.949861 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:04:25 crc kubenswrapper[4688]: I1001 16:04:25.972406 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.024593 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.024636 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsw4d\" (UniqueName: \"kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.024678 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.024908 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.025076 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.025118 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.025162 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.025212 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.126899 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.126974 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.127139 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.127164 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsw4d\" (UniqueName: \"kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.127610 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.128148 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.128572 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.128693 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.128745 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.129054 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.129147 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.135312 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.135626 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.135847 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.144504 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.148159 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsw4d\" (UniqueName: \"kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.168434 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " pod="openstack/glance-default-external-api-0" Oct 01 16:04:26 crc kubenswrapper[4688]: I1001 16:04:26.269261 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:04:27 crc kubenswrapper[4688]: I1001 16:04:27.393863 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" path="/var/lib/kubelet/pods/62e6228f-ade9-4baf-a01f-8d5e19e5836c/volumes" Oct 01 16:04:27 crc kubenswrapper[4688]: I1001 16:04:27.395650 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b38e310-fa47-4eb6-8256-475f992c1efa" path="/var/lib/kubelet/pods/8b38e310-fa47-4eb6-8256-475f992c1efa/volumes" Oct 01 16:04:27 crc kubenswrapper[4688]: I1001 16:04:27.920335 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-77s2p" podUID="62e6228f-ade9-4baf-a01f-8d5e19e5836c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.431351 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.431974 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n687h54ch56fhd8h8fh5fbh5b8h667h75h5b7hcdh695h9fhfch59fh566h6dh5d7h66h68dh67h679h65hf7h5ffh579h678h6bh9h585h64ch5ffq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wsz2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7c9c47cbd9-tpv62_openstack(f7b044e0-4fc7-47c4-a7fd-deda5f608e8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.432981 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.433155 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncdhc8h5c5h5b4h5d4h64h5cfh5cch7fh568hf8hbchb9h64chb6h54fh5b6hcfh587h578h655h665hbh667h5b6h676h694h688hc6hffh56h5cdq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-px8k5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-58744c64d9-c94d2_openstack(1bf07c8b-8b89-473e-a1dc-f008cefe2be4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.433691 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7c9c47cbd9-tpv62" podUID="f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" Oct 01 16:04:31 crc kubenswrapper[4688]: E1001 16:04:31.437657 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-58744c64d9-c94d2" podUID="1bf07c8b-8b89-473e-a1dc-f008cefe2be4" Oct 01 16:04:40 crc kubenswrapper[4688]: I1001 16:04:40.017654 4688 generic.go:334] "Generic (PLEG): container finished" podID="4a989728-c717-4967-a372-93152bd2f064" containerID="8c779883d99658a75782ccb222ae4e75028da3290c76740c0708a162626cab76" exitCode=0 Oct 01 16:04:40 crc kubenswrapper[4688]: I1001 16:04:40.018799 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-flx64" event={"ID":"4a989728-c717-4967-a372-93152bd2f064","Type":"ContainerDied","Data":"8c779883d99658a75782ccb222ae4e75028da3290c76740c0708a162626cab76"} Oct 01 16:04:40 crc kubenswrapper[4688]: I1001 16:04:40.032776 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:40 crc kubenswrapper[4688]: I1001 16:04:40.032867 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:40 crc kubenswrapper[4688]: E1001 16:04:40.769060 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 01 16:04:40 crc kubenswrapper[4688]: E1001 16:04:40.769705 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57bh5d4h677h7fh644h547h679h66ch55ch585h5b4h8fhd8h64ch8h5b4h697h589h65chffh65bh55dh5d6h5f9h89h599hcch98h575h68bhfbh696q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d5wbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(42fa9857-02f6-4a45-9761-93083ddf5104): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:04:41 crc kubenswrapper[4688]: E1001 16:04:41.429980 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 01 16:04:41 crc kubenswrapper[4688]: E1001 16:04:41.430420 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gprb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-bpllf_openstack(8ff202f4-494c-49d4-a228-2e85d6659bb9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:04:41 crc kubenswrapper[4688]: E1001 16:04:41.431748 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-bpllf" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.556609 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.571863 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.578676 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704144 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key\") pod \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704198 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key\") pod \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704268 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704313 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px8k5\" (UniqueName: \"kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5\") pod \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704398 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704442 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts\") pod \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704462 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs\") pod \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704487 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6cxt\" (UniqueName: \"kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704547 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704581 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704606 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data\") pod \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704627 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704667 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts\") pod \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\" (UID: \"1bf07c8b-8b89-473e-a1dc-f008cefe2be4\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704694 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\" (UID: \"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704721 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsz2n\" (UniqueName: \"kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n\") pod \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704748 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs\") pod \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.704786 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data\") pod \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\" (UID: \"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d\") " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.705039 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts" (OuterVolumeSpecName: "scripts") pod "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" (UID: "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.705342 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.705340 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs" (OuterVolumeSpecName: "logs") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.705663 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs" (OuterVolumeSpecName: "logs") pod "1bf07c8b-8b89-473e-a1dc-f008cefe2be4" (UID: "1bf07c8b-8b89-473e-a1dc-f008cefe2be4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.705960 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.706790 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts" (OuterVolumeSpecName: "scripts") pod "1bf07c8b-8b89-473e-a1dc-f008cefe2be4" (UID: "1bf07c8b-8b89-473e-a1dc-f008cefe2be4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.706996 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data" (OuterVolumeSpecName: "config-data") pod "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" (UID: "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.712440 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts" (OuterVolumeSpecName: "scripts") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.713070 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data" (OuterVolumeSpecName: "config-data") pod "1bf07c8b-8b89-473e-a1dc-f008cefe2be4" (UID: "1bf07c8b-8b89-473e-a1dc-f008cefe2be4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.713088 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt" (OuterVolumeSpecName: "kube-api-access-g6cxt") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "kube-api-access-g6cxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.706023 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs" (OuterVolumeSpecName: "logs") pod "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" (UID: "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.716110 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.718044 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1bf07c8b-8b89-473e-a1dc-f008cefe2be4" (UID: "1bf07c8b-8b89-473e-a1dc-f008cefe2be4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.718847 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" (UID: "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.719073 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5" (OuterVolumeSpecName: "kube-api-access-px8k5") pod "1bf07c8b-8b89-473e-a1dc-f008cefe2be4" (UID: "1bf07c8b-8b89-473e-a1dc-f008cefe2be4"). InnerVolumeSpecName "kube-api-access-px8k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.722158 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n" (OuterVolumeSpecName: "kube-api-access-wsz2n") pod "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" (UID: "f7b044e0-4fc7-47c4-a7fd-deda5f608e8d"). InnerVolumeSpecName "kube-api-access-wsz2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.764043 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data" (OuterVolumeSpecName: "config-data") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.766941 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" (UID: "6b37bdf8-ac75-459c-92f9-94e6b9b2d05c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807268 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807299 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px8k5\" (UniqueName: \"kubernetes.io/projected/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-kube-api-access-px8k5\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807313 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807324 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807332 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6cxt\" (UniqueName: \"kubernetes.io/projected/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-kube-api-access-g6cxt\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807340 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807348 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807357 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807365 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807374 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807403 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807413 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsz2n\" (UniqueName: \"kubernetes.io/projected/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-kube-api-access-wsz2n\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807421 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807429 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807437 4688 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.807445 4688 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1bf07c8b-8b89-473e-a1dc-f008cefe2be4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.827768 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 16:04:41 crc kubenswrapper[4688]: I1001 16:04:41.909272 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.036060 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58744c64d9-c94d2" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.036056 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58744c64d9-c94d2" event={"ID":"1bf07c8b-8b89-473e-a1dc-f008cefe2be4","Type":"ContainerDied","Data":"1d5eac5ee2ed41d8a6d447f20d471471b21765350a9dd9e21ac17f4f7b54c1e3"} Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.042164 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b37bdf8-ac75-459c-92f9-94e6b9b2d05c","Type":"ContainerDied","Data":"dcea04d7232475de2626de734db0fd95c235304a8b4978b3eb6a9d2f219854b6"} Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.042247 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.048752 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9c47cbd9-tpv62" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.048847 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9c47cbd9-tpv62" event={"ID":"f7b044e0-4fc7-47c4-a7fd-deda5f608e8d","Type":"ContainerDied","Data":"dd5f262b31c4cc247a56ddeed87f621f8617a8f5fe140424978918fcba76de96"} Oct 01 16:04:42 crc kubenswrapper[4688]: E1001 16:04:42.050393 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-bpllf" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.124174 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.133277 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58744c64d9-c94d2"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.149495 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.164708 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.185208 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:42 crc kubenswrapper[4688]: E1001 16:04:42.185666 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-httpd" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.185688 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-httpd" Oct 01 16:04:42 crc kubenswrapper[4688]: E1001 16:04:42.185708 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-log" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.185718 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-log" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.185876 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-log" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.185897 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" containerName="glance-httpd" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.187001 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.190485 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.191168 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.205560 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.222124 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c9c47cbd9-tpv62"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.227840 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.317947 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318067 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318105 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318135 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318160 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww5wz\" (UniqueName: \"kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318264 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318372 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.318447 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.419971 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420022 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420059 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420078 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420099 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww5wz\" (UniqueName: \"kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420118 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420147 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.420174 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.421638 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.421706 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.421917 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.426278 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.433017 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.433726 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.435062 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.444307 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww5wz\" (UniqueName: \"kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.447499 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.514840 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:42 crc kubenswrapper[4688]: I1001 16:04:42.989063 4688 scope.go:117] "RemoveContainer" containerID="fb0f2877979d065e9427d7f77859cc896504e685b8d7529195c0334076f38602" Oct 01 16:04:43 crc kubenswrapper[4688]: E1001 16:04:43.039382 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 01 16:04:43 crc kubenswrapper[4688]: E1001 16:04:43.039508 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgsvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-bqw9p_openstack(81b3c008-7340-436e-873b-aa0d41588c6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:04:43 crc kubenswrapper[4688]: E1001 16:04:43.041013 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-bqw9p" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.066786 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-flx64" event={"ID":"4a989728-c717-4967-a372-93152bd2f064","Type":"ContainerDied","Data":"970c8c2d9e9c586d88e639a9f8ca2f73cb1fa621ef664b649c343a5f3766000d"} Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.067014 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970c8c2d9e9c586d88e639a9f8ca2f73cb1fa621ef664b649c343a5f3766000d" Oct 01 16:04:43 crc kubenswrapper[4688]: E1001 16:04:43.068728 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-bqw9p" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.127499 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.239625 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw95n\" (UniqueName: \"kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n\") pod \"4a989728-c717-4967-a372-93152bd2f064\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.239738 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle\") pod \"4a989728-c717-4967-a372-93152bd2f064\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.239864 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config\") pod \"4a989728-c717-4967-a372-93152bd2f064\" (UID: \"4a989728-c717-4967-a372-93152bd2f064\") " Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.244628 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n" (OuterVolumeSpecName: "kube-api-access-zw95n") pod "4a989728-c717-4967-a372-93152bd2f064" (UID: "4a989728-c717-4967-a372-93152bd2f064"). InnerVolumeSpecName "kube-api-access-zw95n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.266816 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a989728-c717-4967-a372-93152bd2f064" (UID: "4a989728-c717-4967-a372-93152bd2f064"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.271762 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config" (OuterVolumeSpecName: "config") pod "4a989728-c717-4967-a372-93152bd2f064" (UID: "4a989728-c717-4967-a372-93152bd2f064"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.341686 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.341722 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw95n\" (UniqueName: \"kubernetes.io/projected/4a989728-c717-4967-a372-93152bd2f064-kube-api-access-zw95n\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.341744 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a989728-c717-4967-a372-93152bd2f064-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.392871 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf07c8b-8b89-473e-a1dc-f008cefe2be4" path="/var/lib/kubelet/pods/1bf07c8b-8b89-473e-a1dc-f008cefe2be4/volumes" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.393300 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b37bdf8-ac75-459c-92f9-94e6b9b2d05c" path="/var/lib/kubelet/pods/6b37bdf8-ac75-459c-92f9-94e6b9b2d05c/volumes" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.394052 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7b044e0-4fc7-47c4-a7fd-deda5f608e8d" path="/var/lib/kubelet/pods/f7b044e0-4fc7-47c4-a7fd-deda5f608e8d/volumes" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.541372 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-555656bf5b-xlntq"] Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.575680 4688 scope.go:117] "RemoveContainer" containerID="e7b9aa1bdf9f1fb38248175acc41dd1be51e9764e75233db32aa665d91636bbc" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.622464 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-82bf8"] Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.708928 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:04:43 crc kubenswrapper[4688]: W1001 16:04:43.720562 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf94f981e_54a9_4bca_ad26_518b47ca0fb3.slice/crio-7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e WatchSource:0}: Error finding container 7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e: Status 404 returned error can't find the container with id 7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e Oct 01 16:04:43 crc kubenswrapper[4688]: W1001 16:04:43.736877 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf24d7e9c_1e3f_4a7b_b88e_234b98f4ac2e.slice/crio-8c0c2ff4a3dd5be834a2e028a22c76a76b5ef50aa7d3a8867165c97af80d3210 WatchSource:0}: Error finding container 8c0c2ff4a3dd5be834a2e028a22c76a76b5ef50aa7d3a8867165c97af80d3210: Status 404 returned error can't find the container with id 8c0c2ff4a3dd5be834a2e028a22c76a76b5ef50aa7d3a8867165c97af80d3210 Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.740951 4688 scope.go:117] "RemoveContainer" containerID="9c0aa2322502dbe9c71c6829902f6b42366ba368d03602bca5cda9e653c79544" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.787696 4688 scope.go:117] "RemoveContainer" containerID="e371a796da46d1f12192e2cb38d4b16d78038d5ff715684faa42c5c40ba709ca" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.832694 4688 scope.go:117] "RemoveContainer" containerID="3162bf0e25e7d61e27cf4f11caf9dbb1fe4462353ddad47ee4284f48988db10c" Oct 01 16:04:43 crc kubenswrapper[4688]: I1001 16:04:43.970806 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:04:43 crc kubenswrapper[4688]: W1001 16:04:43.980282 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85db256a_98d5_4bca_82ac_098d49b243e8.slice/crio-0b91ae9614216b0bfe30f98a0e38341b10439941cfab805630039fa06af1b15e WatchSource:0}: Error finding container 0b91ae9614216b0bfe30f98a0e38341b10439941cfab805630039fa06af1b15e: Status 404 returned error can't find the container with id 0b91ae9614216b0bfe30f98a0e38341b10439941cfab805630039fa06af1b15e Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.104953 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerStarted","Data":"8c0c2ff4a3dd5be834a2e028a22c76a76b5ef50aa7d3a8867165c97af80d3210"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.107384 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerStarted","Data":"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.109239 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82bf8" event={"ID":"f94f981e-54a9-4bca-ad26-518b47ca0fb3","Type":"ContainerStarted","Data":"7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.119043 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-555656bf5b-xlntq" event={"ID":"37923e92-dbcc-41a1-8d2f-89d8de59959e","Type":"ContainerStarted","Data":"14ba7a3e43b554c23d312880e1e92140525d8bc75b34d855ecbca11db6f90657"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.119119 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-555656bf5b-xlntq" event={"ID":"37923e92-dbcc-41a1-8d2f-89d8de59959e","Type":"ContainerStarted","Data":"9c64cc31772477cc2485832e0a18bed6018c8372d0db45f994bb2b3df7af1a9b"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.120837 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ntpwg" event={"ID":"26e68b63-d053-412d-95e8-0e5f10bffc45","Type":"ContainerStarted","Data":"b5de4ddd09f876e9803f292b21becf10b4aa56aab7370c7d20878f1a40ecc847"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.138743 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ntpwg" podStartSLOduration=4.828256628 podStartE2EDuration="39.138696671s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.120813179 +0000 UTC m=+1036.471453141" lastFinishedPulling="2025-10-01 16:04:41.431253222 +0000 UTC m=+1070.781893184" observedRunningTime="2025-10-01 16:04:44.135661926 +0000 UTC m=+1073.486301888" watchObservedRunningTime="2025-10-01 16:04:44.138696671 +0000 UTC m=+1073.489336633" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.139835 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerStarted","Data":"0b91ae9614216b0bfe30f98a0e38341b10439941cfab805630039fa06af1b15e"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.144252 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-flx64" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.145106 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerStarted","Data":"2b7f7e054af4f411521fe37a0501664525292ea06e93ed75734516903533122c"} Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.196437 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.429333 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:04:44 crc kubenswrapper[4688]: E1001 16:04:44.442680 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a989728-c717-4967-a372-93152bd2f064" containerName="neutron-db-sync" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.442716 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a989728-c717-4967-a372-93152bd2f064" containerName="neutron-db-sync" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.442958 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a989728-c717-4967-a372-93152bd2f064" containerName="neutron-db-sync" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.444151 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.481056 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.558561 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.559903 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.567015 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.567190 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.586424 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qzsnx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.586698 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588593 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588644 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lfh9\" (UniqueName: \"kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588671 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588691 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588722 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.588781 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.621579 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690377 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690428 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lfh9\" (UniqueName: \"kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690456 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690473 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690496 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690516 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690554 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690573 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690602 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690661 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.690684 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw5cn\" (UniqueName: \"kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.691409 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.691830 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.691931 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.692448 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.692950 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.716298 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lfh9\" (UniqueName: \"kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9\") pod \"dnsmasq-dns-6b7b667979-k7bhx\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.792825 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.792909 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.792935 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.793035 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.793069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw5cn\" (UniqueName: \"kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.802004 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.804728 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.807962 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.826173 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.834654 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw5cn\" (UniqueName: \"kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn\") pod \"neutron-f8467544d-9fz8w\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.911896 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:44 crc kubenswrapper[4688]: I1001 16:04:44.953192 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.226654 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerStarted","Data":"13d7ca8afb588a2930fb85629b299a193fbcd9d5cd0e456cc1c1a02defa60648"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.228612 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-555656bf5b-xlntq" event={"ID":"37923e92-dbcc-41a1-8d2f-89d8de59959e","Type":"ContainerStarted","Data":"674bcb3920fc3555ff6521c9241431dd30559cff036703357069642077847a95"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.255015 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-555656bf5b-xlntq" podStartSLOduration=27.254997772 podStartE2EDuration="27.254997772s" podCreationTimestamp="2025-10-01 16:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:45.251315989 +0000 UTC m=+1074.601955951" watchObservedRunningTime="2025-10-01 16:04:45.254997772 +0000 UTC m=+1074.605637724" Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.280064 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerStarted","Data":"db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.280104 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerStarted","Data":"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.296809 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerStarted","Data":"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.296954 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5889fcf659-ptwxd" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon-log" containerID="cri-o://93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" gracePeriod=30 Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.297210 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5889fcf659-ptwxd" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon" containerID="cri-o://047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" gracePeriod=30 Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.321265 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-654b94b7dd-g2k9q" podStartSLOduration=27.321250514 podStartE2EDuration="27.321250514s" podCreationTimestamp="2025-10-01 16:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:45.318041375 +0000 UTC m=+1074.668681327" watchObservedRunningTime="2025-10-01 16:04:45.321250514 +0000 UTC m=+1074.671890476" Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.326940 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerStarted","Data":"e7a26a406d05cbe87f75cf5049289106ab60f10b02d02313f13255906a60ef2f"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.332387 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82bf8" event={"ID":"f94f981e-54a9-4bca-ad26-518b47ca0fb3","Type":"ContainerStarted","Data":"0d0979eba270753cd514b797174ad6d89b0308dcdacd47698c0539f0061871cb"} Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.351954 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5889fcf659-ptwxd" podStartSLOduration=6.477809825 podStartE2EDuration="40.351935298s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.583847743 +0000 UTC m=+1036.934487705" lastFinishedPulling="2025-10-01 16:04:41.457973216 +0000 UTC m=+1070.808613178" observedRunningTime="2025-10-01 16:04:45.341662222 +0000 UTC m=+1074.692302184" watchObservedRunningTime="2025-10-01 16:04:45.351935298 +0000 UTC m=+1074.702575260" Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.454846 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-82bf8" podStartSLOduration=26.45482422 podStartE2EDuration="26.45482422s" podCreationTimestamp="2025-10-01 16:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:45.363707365 +0000 UTC m=+1074.714347327" watchObservedRunningTime="2025-10-01 16:04:45.45482422 +0000 UTC m=+1074.805464182" Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.656817 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:04:45 crc kubenswrapper[4688]: W1001 16:04:45.681966 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef2b662d_30ed_497e_84e2_a4cdde464337.slice/crio-f666241b4207d73a8c49fc648bc899f0ddd52329c1387da1b4b355a31cde9ea4 WatchSource:0}: Error finding container f666241b4207d73a8c49fc648bc899f0ddd52329c1387da1b4b355a31cde9ea4: Status 404 returned error can't find the container with id f666241b4207d73a8c49fc648bc899f0ddd52329c1387da1b4b355a31cde9ea4 Oct 01 16:04:45 crc kubenswrapper[4688]: I1001 16:04:45.934847 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.045144 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.346260 4688 generic.go:334] "Generic (PLEG): container finished" podID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerID="fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527" exitCode=0 Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.346312 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" event={"ID":"ef2b662d-30ed-497e-84e2-a4cdde464337","Type":"ContainerDied","Data":"fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527"} Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.346361 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" event={"ID":"ef2b662d-30ed-497e-84e2-a4cdde464337","Type":"ContainerStarted","Data":"f666241b4207d73a8c49fc648bc899f0ddd52329c1387da1b4b355a31cde9ea4"} Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.357983 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerStarted","Data":"46257c1c7425d0ab02e789a29a47d108581911b07f7b67fed4e7e9bd27c4d1ef"} Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.362972 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerStarted","Data":"669ce4b84dfa8d9deb9d930ec89a60575ab320d2060fc2071b9e14a15b19aa6e"} Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.369801 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerStarted","Data":"d4c516a2d85e1b38d9b2435307a28c8f856bc9f20bfd888e02f60137f262ee6c"} Oct 01 16:04:46 crc kubenswrapper[4688]: I1001 16:04:46.418245 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=21.418210967 podStartE2EDuration="21.418210967s" podCreationTimestamp="2025-10-01 16:04:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:46.415831161 +0000 UTC m=+1075.766471123" watchObservedRunningTime="2025-10-01 16:04:46.418210967 +0000 UTC m=+1075.768850919" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.389863 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerStarted","Data":"73abce7839d4c71f72b092bcafbba0a7ec6f26e945f08b62e9c618f93d3b4171"} Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.390329 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.390340 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerStarted","Data":"c00340087753728353eab21fd7726560d201e35cc7849e4943ecbf0cdf0504d0"} Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.394372 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" event={"ID":"ef2b662d-30ed-497e-84e2-a4cdde464337","Type":"ContainerStarted","Data":"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a"} Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.395294 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.400058 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerStarted","Data":"60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f"} Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.408935 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f8467544d-9fz8w" podStartSLOduration=3.408913943 podStartE2EDuration="3.408913943s" podCreationTimestamp="2025-10-01 16:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:47.403304117 +0000 UTC m=+1076.753944099" watchObservedRunningTime="2025-10-01 16:04:47.408913943 +0000 UTC m=+1076.759553905" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.442172 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.442154438 podStartE2EDuration="5.442154438s" podCreationTimestamp="2025-10-01 16:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:47.437016525 +0000 UTC m=+1076.787656507" watchObservedRunningTime="2025-10-01 16:04:47.442154438 +0000 UTC m=+1076.792794400" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.475998 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" podStartSLOduration=3.475979219 podStartE2EDuration="3.475979219s" podCreationTimestamp="2025-10-01 16:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:47.473053237 +0000 UTC m=+1076.823693209" watchObservedRunningTime="2025-10-01 16:04:47.475979219 +0000 UTC m=+1076.826619181" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.513558 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c946d4bf9-l8zds"] Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.514885 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.535513 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.535979 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.550305 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c946d4bf9-l8zds"] Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.566747 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-public-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567048 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jddb7\" (UniqueName: \"kubernetes.io/projected/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-kube-api-access-jddb7\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567103 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567125 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-combined-ca-bundle\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567148 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-httpd-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567171 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-internal-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.567191 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-ovndb-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668189 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-internal-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668234 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-ovndb-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668331 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-public-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jddb7\" (UniqueName: \"kubernetes.io/projected/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-kube-api-access-jddb7\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668423 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668448 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-combined-ca-bundle\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.668469 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-httpd-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.683071 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-combined-ca-bundle\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.683466 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-internal-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.683767 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-ovndb-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.698483 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.698683 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-public-tls-certs\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.721409 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-httpd-config\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.721741 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jddb7\" (UniqueName: \"kubernetes.io/projected/5766bcc5-cf95-400d-a634-c2e0ad4fe3b9-kube-api-access-jddb7\") pod \"neutron-c946d4bf9-l8zds\" (UID: \"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9\") " pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:47 crc kubenswrapper[4688]: I1001 16:04:47.863327 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:48 crc kubenswrapper[4688]: I1001 16:04:48.538808 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:48 crc kubenswrapper[4688]: I1001 16:04:48.539159 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:04:48 crc kubenswrapper[4688]: I1001 16:04:48.588451 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:48 crc kubenswrapper[4688]: I1001 16:04:48.588551 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:04:48 crc kubenswrapper[4688]: I1001 16:04:48.606445 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c946d4bf9-l8zds"] Oct 01 16:04:49 crc kubenswrapper[4688]: I1001 16:04:49.432514 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c946d4bf9-l8zds" event={"ID":"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9","Type":"ContainerStarted","Data":"c7f9a07c0d377599d54ada2ee9471a1675717019047f30f98b2d7693b672d2a1"} Oct 01 16:04:49 crc kubenswrapper[4688]: I1001 16:04:49.433062 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c946d4bf9-l8zds" event={"ID":"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9","Type":"ContainerStarted","Data":"6fd1fd06c73fda607ce32aae6eacedc863f4295706ab4103742813382aa7b349"} Oct 01 16:04:51 crc kubenswrapper[4688]: I1001 16:04:51.480179 4688 generic.go:334] "Generic (PLEG): container finished" podID="26e68b63-d053-412d-95e8-0e5f10bffc45" containerID="b5de4ddd09f876e9803f292b21becf10b4aa56aab7370c7d20878f1a40ecc847" exitCode=0 Oct 01 16:04:51 crc kubenswrapper[4688]: I1001 16:04:51.481590 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ntpwg" event={"ID":"26e68b63-d053-412d-95e8-0e5f10bffc45","Type":"ContainerDied","Data":"b5de4ddd09f876e9803f292b21becf10b4aa56aab7370c7d20878f1a40ecc847"} Oct 01 16:04:52 crc kubenswrapper[4688]: I1001 16:04:52.515918 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:52 crc kubenswrapper[4688]: I1001 16:04:52.516222 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:52 crc kubenswrapper[4688]: I1001 16:04:52.584417 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:52 crc kubenswrapper[4688]: I1001 16:04:52.585707 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.500945 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ntpwg" event={"ID":"26e68b63-d053-412d-95e8-0e5f10bffc45","Type":"ContainerDied","Data":"2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7"} Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.501196 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a7492c42c35eed2eba95aec8c8415cd16efafd61410b58ffc56c030787e09f7" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.502913 4688 generic.go:334] "Generic (PLEG): container finished" podID="f94f981e-54a9-4bca-ad26-518b47ca0fb3" containerID="0d0979eba270753cd514b797174ad6d89b0308dcdacd47698c0539f0061871cb" exitCode=0 Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.502943 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82bf8" event={"ID":"f94f981e-54a9-4bca-ad26-518b47ca0fb3","Type":"ContainerDied","Data":"0d0979eba270753cd514b797174ad6d89b0308dcdacd47698c0539f0061871cb"} Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.504572 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.504600 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.532071 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.634572 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts\") pod \"26e68b63-d053-412d-95e8-0e5f10bffc45\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.635165 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle\") pod \"26e68b63-d053-412d-95e8-0e5f10bffc45\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.635350 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brgzp\" (UniqueName: \"kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp\") pod \"26e68b63-d053-412d-95e8-0e5f10bffc45\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.635490 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data\") pod \"26e68b63-d053-412d-95e8-0e5f10bffc45\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.635966 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs\") pod \"26e68b63-d053-412d-95e8-0e5f10bffc45\" (UID: \"26e68b63-d053-412d-95e8-0e5f10bffc45\") " Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.636334 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs" (OuterVolumeSpecName: "logs") pod "26e68b63-d053-412d-95e8-0e5f10bffc45" (UID: "26e68b63-d053-412d-95e8-0e5f10bffc45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.644267 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp" (OuterVolumeSpecName: "kube-api-access-brgzp") pod "26e68b63-d053-412d-95e8-0e5f10bffc45" (UID: "26e68b63-d053-412d-95e8-0e5f10bffc45"). InnerVolumeSpecName "kube-api-access-brgzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.654048 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts" (OuterVolumeSpecName: "scripts") pod "26e68b63-d053-412d-95e8-0e5f10bffc45" (UID: "26e68b63-d053-412d-95e8-0e5f10bffc45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.671672 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data" (OuterVolumeSpecName: "config-data") pod "26e68b63-d053-412d-95e8-0e5f10bffc45" (UID: "26e68b63-d053-412d-95e8-0e5f10bffc45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.691397 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26e68b63-d053-412d-95e8-0e5f10bffc45" (UID: "26e68b63-d053-412d-95e8-0e5f10bffc45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.739234 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.739269 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.739283 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brgzp\" (UniqueName: \"kubernetes.io/projected/26e68b63-d053-412d-95e8-0e5f10bffc45-kube-api-access-brgzp\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.739292 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e68b63-d053-412d-95e8-0e5f10bffc45-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:53 crc kubenswrapper[4688]: I1001 16:04:53.739301 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e68b63-d053-412d-95e8-0e5f10bffc45-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.514674 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerStarted","Data":"77433936d791e20959e176cedf7427bd592cbdef3cc6723d0c7b18814b89c22d"} Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.516935 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c946d4bf9-l8zds" event={"ID":"5766bcc5-cf95-400d-a634-c2e0ad4fe3b9","Type":"ContainerStarted","Data":"e1b2df1e5661fdf97f9debe6feca3be69dd9a5273a4c0447d4d2df44326453ac"} Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.517073 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ntpwg" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.551003 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c946d4bf9-l8zds" podStartSLOduration=7.550981792 podStartE2EDuration="7.550981792s" podCreationTimestamp="2025-10-01 16:04:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:54.536561861 +0000 UTC m=+1083.887201823" watchObservedRunningTime="2025-10-01 16:04:54.550981792 +0000 UTC m=+1083.901621754" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.740630 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-76d4fdc768-gl782"] Oct 01 16:04:54 crc kubenswrapper[4688]: E1001 16:04:54.741917 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e68b63-d053-412d-95e8-0e5f10bffc45" containerName="placement-db-sync" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.741930 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e68b63-d053-412d-95e8-0e5f10bffc45" containerName="placement-db-sync" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.742088 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e68b63-d053-412d-95e8-0e5f10bffc45" containerName="placement-db-sync" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.742975 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.750717 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.750894 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.751047 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.751160 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.751269 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-955dk" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.782381 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-76d4fdc768-gl782"] Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848175 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-internal-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848348 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb2d5\" (UniqueName: \"kubernetes.io/projected/d97f10ed-cfae-4409-b6fd-ae047f584ac3-kube-api-access-hb2d5\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848431 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d97f10ed-cfae-4409-b6fd-ae047f584ac3-logs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848505 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-config-data\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848587 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-public-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848820 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-scripts\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.848906 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-combined-ca-bundle\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.913742 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.950755 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-public-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951054 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-scripts\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951085 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-combined-ca-bundle\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951123 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-internal-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951166 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb2d5\" (UniqueName: \"kubernetes.io/projected/d97f10ed-cfae-4409-b6fd-ae047f584ac3-kube-api-access-hb2d5\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951202 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d97f10ed-cfae-4409-b6fd-ae047f584ac3-logs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.951220 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-config-data\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.953032 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d97f10ed-cfae-4409-b6fd-ae047f584ac3-logs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.957169 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-internal-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.960075 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-config-data\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.967368 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-scripts\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.976206 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-public-tls-certs\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:54 crc kubenswrapper[4688]: I1001 16:04:54.984433 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f10ed-cfae-4409-b6fd-ae047f584ac3-combined-ca-bundle\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.020112 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb2d5\" (UniqueName: \"kubernetes.io/projected/d97f10ed-cfae-4409-b6fd-ae047f584ac3-kube-api-access-hb2d5\") pod \"placement-76d4fdc768-gl782\" (UID: \"d97f10ed-cfae-4409-b6fd-ae047f584ac3\") " pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.048737 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.048953 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="dnsmasq-dns" containerID="cri-o://1f14ba0ef31a4de3533a673ab57ef7043aeb05057ada91913157bcbf4345a8a1" gracePeriod=10 Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.058671 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.058728 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.084802 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.093053 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160754 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160815 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p69v4\" (UniqueName: \"kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160895 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160939 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160968 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.160993 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys\") pod \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\" (UID: \"f94f981e-54a9-4bca-ad26-518b47ca0fb3\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.178042 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.178599 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.178231 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4" (OuterVolumeSpecName: "kube-api-access-p69v4") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "kube-api-access-p69v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.179151 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts" (OuterVolumeSpecName: "scripts") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.212894 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data" (OuterVolumeSpecName: "config-data") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.214954 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f94f981e-54a9-4bca-ad26-518b47ca0fb3" (UID: "f94f981e-54a9-4bca-ad26-518b47ca0fb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263561 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263587 4688 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263596 4688 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263604 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p69v4\" (UniqueName: \"kubernetes.io/projected/f94f981e-54a9-4bca-ad26-518b47ca0fb3-kube-api-access-p69v4\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263615 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.263623 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94f981e-54a9-4bca-ad26-518b47ca0fb3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.536259 4688 generic.go:334] "Generic (PLEG): container finished" podID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerID="1f14ba0ef31a4de3533a673ab57ef7043aeb05057ada91913157bcbf4345a8a1" exitCode=0 Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.536554 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" event={"ID":"7e1915c3-1db9-4c2f-8011-df62c28827af","Type":"ContainerDied","Data":"1f14ba0ef31a4de3533a673ab57ef7043aeb05057ada91913157bcbf4345a8a1"} Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.539733 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bpllf" event={"ID":"8ff202f4-494c-49d4-a228-2e85d6659bb9","Type":"ContainerStarted","Data":"d70829f518e96ec873ead8353a810739e03111c685e2196b634ec07a720ded11"} Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.574839 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.574863 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.575359 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-82bf8" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.575770 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-82bf8" event={"ID":"f94f981e-54a9-4bca-ad26-518b47ca0fb3","Type":"ContainerDied","Data":"7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e"} Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.575789 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b0ffb3208429e08b392cbe9b705fe927924c29b265fe3c927166b75138dbb8e" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.576251 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.584499 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-bpllf" podStartSLOduration=2.700724384 podStartE2EDuration="50.58447692s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.052821212 +0000 UTC m=+1036.403461174" lastFinishedPulling="2025-10-01 16:04:54.936573748 +0000 UTC m=+1084.287213710" observedRunningTime="2025-10-01 16:04:55.566095509 +0000 UTC m=+1084.916735471" watchObservedRunningTime="2025-10-01 16:04:55.58447692 +0000 UTC m=+1084.935116882" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.694070 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-587df57d87-hlpwt"] Oct 01 16:04:55 crc kubenswrapper[4688]: E1001 16:04:55.694486 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94f981e-54a9-4bca-ad26-518b47ca0fb3" containerName="keystone-bootstrap" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.694562 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94f981e-54a9-4bca-ad26-518b47ca0fb3" containerName="keystone-bootstrap" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.694815 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94f981e-54a9-4bca-ad26-518b47ca0fb3" containerName="keystone-bootstrap" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.695731 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.717139 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.717594 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.717793 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.717996 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.718169 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.718391 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zcpfl" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.718922 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-76d4fdc768-gl782"] Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.744931 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-587df57d87-hlpwt"] Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.851203 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886390 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-internal-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886453 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-fernet-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886494 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-scripts\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886514 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-public-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886558 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-config-data\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886644 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rhx\" (UniqueName: \"kubernetes.io/projected/8a94d646-b2be-4ab1-89ff-da8aa30591f6-kube-api-access-g6rhx\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886680 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-credential-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.886709 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-combined-ca-bundle\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.989563 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.989866 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.989890 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7fpq\" (UniqueName: \"kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.989931 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990019 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990064 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0\") pod \"7e1915c3-1db9-4c2f-8011-df62c28827af\" (UID: \"7e1915c3-1db9-4c2f-8011-df62c28827af\") " Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990341 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-internal-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990374 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-fernet-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990405 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-scripts\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990423 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-public-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990443 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-config-data\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990511 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rhx\" (UniqueName: \"kubernetes.io/projected/8a94d646-b2be-4ab1-89ff-da8aa30591f6-kube-api-access-g6rhx\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990559 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-credential-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.990584 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-combined-ca-bundle\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:55 crc kubenswrapper[4688]: I1001 16:04:55.994819 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-combined-ca-bundle\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.004418 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-config-data\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.007186 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq" (OuterVolumeSpecName: "kube-api-access-z7fpq") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "kube-api-access-z7fpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.009057 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-credential-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.012863 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-public-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.015035 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-scripts\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.015431 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-fernet-keys\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.016984 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rhx\" (UniqueName: \"kubernetes.io/projected/8a94d646-b2be-4ab1-89ff-da8aa30591f6-kube-api-access-g6rhx\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.023834 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a94d646-b2be-4ab1-89ff-da8aa30591f6-internal-tls-certs\") pod \"keystone-587df57d87-hlpwt\" (UID: \"8a94d646-b2be-4ab1-89ff-da8aa30591f6\") " pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.082623 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.091976 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.092009 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7fpq\" (UniqueName: \"kubernetes.io/projected/7e1915c3-1db9-4c2f-8011-df62c28827af-kube-api-access-z7fpq\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.119118 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.145714 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config" (OuterVolumeSpecName: "config") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.149660 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.159046 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e1915c3-1db9-4c2f-8011-df62c28827af" (UID: "7e1915c3-1db9-4c2f-8011-df62c28827af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.191604 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.193010 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.193023 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.193035 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.193045 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e1915c3-1db9-4c2f-8011-df62c28827af-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.271744 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.271782 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.271792 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.271804 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.354372 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.439472 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.597551 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" event={"ID":"7e1915c3-1db9-4c2f-8011-df62c28827af","Type":"ContainerDied","Data":"58523a79bc353c3b058ae84a9c282ad2bf651b68a72d02eb01f27c86f4ac00ae"} Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.597601 4688 scope.go:117] "RemoveContainer" containerID="1f14ba0ef31a4de3533a673ab57ef7043aeb05057ada91913157bcbf4345a8a1" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.597735 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-r48br" Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.640570 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-76d4fdc768-gl782" event={"ID":"d97f10ed-cfae-4409-b6fd-ae047f584ac3","Type":"ContainerStarted","Data":"1c02bd9f4f54b2c2861382b8c9bbfe193cfa492a8ee18468c8bba886b437a7cc"} Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.640623 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-76d4fdc768-gl782" event={"ID":"d97f10ed-cfae-4409-b6fd-ae047f584ac3","Type":"ContainerStarted","Data":"3a3b781c011272671427b6fe22f7ba1d14dcebf67079db1002e28b3972656e99"} Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.652696 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.675095 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-r48br"] Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.818670 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-587df57d87-hlpwt"] Oct 01 16:04:56 crc kubenswrapper[4688]: I1001 16:04:56.929216 4688 scope.go:117] "RemoveContainer" containerID="d68f0020f5402bff3d6afb5d51c93e9f48ae69749b685812a5dea4b67412f4b0" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.394818 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" path="/var/lib/kubelet/pods/7e1915c3-1db9-4c2f-8011-df62c28827af/volumes" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.665486 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-587df57d87-hlpwt" event={"ID":"8a94d646-b2be-4ab1-89ff-da8aa30591f6","Type":"ContainerStarted","Data":"d4cfc339a8281993180f6d7baaf25f48e127b62e632c9d9e7273d6442f276cb1"} Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.665823 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-587df57d87-hlpwt" event={"ID":"8a94d646-b2be-4ab1-89ff-da8aa30591f6","Type":"ContainerStarted","Data":"ad40de0e74091513ad38e749c4396145450f839d78bb8d141421e53b10f8c04e"} Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.665845 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.668178 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqw9p" event={"ID":"81b3c008-7340-436e-873b-aa0d41588c6c","Type":"ContainerStarted","Data":"287025f6387c3e20194c2bf8af4a31a4fe8ed192538b47574fd21a7d50bfdff3"} Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.681652 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-76d4fdc768-gl782" event={"ID":"d97f10ed-cfae-4409-b6fd-ae047f584ac3","Type":"ContainerStarted","Data":"57f2f82c098acd06e8620167810fc81afcb8ea15025bfaf14820f841017a2875"} Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.681738 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.681944 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.689863 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-587df57d87-hlpwt" podStartSLOduration=2.689840852 podStartE2EDuration="2.689840852s" podCreationTimestamp="2025-10-01 16:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:57.685788439 +0000 UTC m=+1087.036428401" watchObservedRunningTime="2025-10-01 16:04:57.689840852 +0000 UTC m=+1087.040480814" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.721566 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-76d4fdc768-gl782" podStartSLOduration=3.721548604 podStartE2EDuration="3.721548604s" podCreationTimestamp="2025-10-01 16:04:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:04:57.719205759 +0000 UTC m=+1087.069845721" watchObservedRunningTime="2025-10-01 16:04:57.721548604 +0000 UTC m=+1087.072188566" Oct 01 16:04:57 crc kubenswrapper[4688]: I1001 16:04:57.739743 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-bqw9p" podStartSLOduration=3.737118019 podStartE2EDuration="52.73972663s" podCreationTimestamp="2025-10-01 16:04:05 +0000 UTC" firstStartedPulling="2025-10-01 16:04:07.079975824 +0000 UTC m=+1036.430615786" lastFinishedPulling="2025-10-01 16:04:56.082584435 +0000 UTC m=+1085.433224397" observedRunningTime="2025-10-01 16:04:57.739123383 +0000 UTC m=+1087.089763345" watchObservedRunningTime="2025-10-01 16:04:57.73972663 +0000 UTC m=+1087.090366592" Oct 01 16:04:58 crc kubenswrapper[4688]: I1001 16:04:58.540155 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:04:58 crc kubenswrapper[4688]: I1001 16:04:58.589280 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Oct 01 16:04:59 crc kubenswrapper[4688]: I1001 16:04:59.674282 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:04:59 crc kubenswrapper[4688]: I1001 16:04:59.674376 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:04:59 crc kubenswrapper[4688]: I1001 16:04:59.682069 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:00 crc kubenswrapper[4688]: I1001 16:05:00.115542 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:05:00 crc kubenswrapper[4688]: I1001 16:05:00.116067 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:05:00 crc kubenswrapper[4688]: I1001 16:05:00.120743 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:05:02 crc kubenswrapper[4688]: I1001 16:05:02.743837 4688 generic.go:334] "Generic (PLEG): container finished" podID="8ff202f4-494c-49d4-a228-2e85d6659bb9" containerID="d70829f518e96ec873ead8353a810739e03111c685e2196b634ec07a720ded11" exitCode=0 Oct 01 16:05:02 crc kubenswrapper[4688]: I1001 16:05:02.744170 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bpllf" event={"ID":"8ff202f4-494c-49d4-a228-2e85d6659bb9","Type":"ContainerDied","Data":"d70829f518e96ec873ead8353a810739e03111c685e2196b634ec07a720ded11"} Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.537751 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bpllf" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.608105 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle\") pod \"8ff202f4-494c-49d4-a228-2e85d6659bb9\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.608201 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data\") pod \"8ff202f4-494c-49d4-a228-2e85d6659bb9\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.608221 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gprb5\" (UniqueName: \"kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5\") pod \"8ff202f4-494c-49d4-a228-2e85d6659bb9\" (UID: \"8ff202f4-494c-49d4-a228-2e85d6659bb9\") " Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.613399 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8ff202f4-494c-49d4-a228-2e85d6659bb9" (UID: "8ff202f4-494c-49d4-a228-2e85d6659bb9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.629176 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5" (OuterVolumeSpecName: "kube-api-access-gprb5") pod "8ff202f4-494c-49d4-a228-2e85d6659bb9" (UID: "8ff202f4-494c-49d4-a228-2e85d6659bb9"). InnerVolumeSpecName "kube-api-access-gprb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.658596 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ff202f4-494c-49d4-a228-2e85d6659bb9" (UID: "8ff202f4-494c-49d4-a228-2e85d6659bb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.710511 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.710574 4688 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ff202f4-494c-49d4-a228-2e85d6659bb9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.710587 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gprb5\" (UniqueName: \"kubernetes.io/projected/8ff202f4-494c-49d4-a228-2e85d6659bb9-kube-api-access-gprb5\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.781234 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bpllf" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.781237 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bpllf" event={"ID":"8ff202f4-494c-49d4-a228-2e85d6659bb9","Type":"ContainerDied","Data":"6d2f41598d0ec8132a19ce85408af5e6624a4071db79cc403cbfde8d9e9d9358"} Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.781299 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d2f41598d0ec8132a19ce85408af5e6624a4071db79cc403cbfde8d9e9d9358" Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.784207 4688 generic.go:334] "Generic (PLEG): container finished" podID="81b3c008-7340-436e-873b-aa0d41588c6c" containerID="287025f6387c3e20194c2bf8af4a31a4fe8ed192538b47574fd21a7d50bfdff3" exitCode=0 Oct 01 16:05:06 crc kubenswrapper[4688]: I1001 16:05:06.784241 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqw9p" event={"ID":"81b3c008-7340-436e-873b-aa0d41588c6c","Type":"ContainerDied","Data":"287025f6387c3e20194c2bf8af4a31a4fe8ed192538b47574fd21a7d50bfdff3"} Oct 01 16:05:07 crc kubenswrapper[4688]: E1001 16:05:07.294558 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.793650 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerStarted","Data":"d4a9a43158e938fbb69326381e5b182bdb47e0f1e8f611b017e1bee9b41d1c0c"} Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.793846 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="ceilometer-notification-agent" containerID="cri-o://2b7f7e054af4f411521fe37a0501664525292ea06e93ed75734516903533122c" gracePeriod=30 Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.793894 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="proxy-httpd" containerID="cri-o://d4a9a43158e938fbb69326381e5b182bdb47e0f1e8f611b017e1bee9b41d1c0c" gracePeriod=30 Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.793936 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="sg-core" containerID="cri-o://77433936d791e20959e176cedf7427bd592cbdef3cc6723d0c7b18814b89c22d" gracePeriod=30 Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.927616 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-df57dbd74-8gcrt"] Oct 01 16:05:07 crc kubenswrapper[4688]: E1001 16:05:07.927997 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" containerName="barbican-db-sync" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.928013 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" containerName="barbican-db-sync" Oct 01 16:05:07 crc kubenswrapper[4688]: E1001 16:05:07.928034 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="init" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.928040 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="init" Oct 01 16:05:07 crc kubenswrapper[4688]: E1001 16:05:07.928057 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="dnsmasq-dns" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.928065 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="dnsmasq-dns" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.932339 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e1915c3-1db9-4c2f-8011-df62c28827af" containerName="dnsmasq-dns" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.932419 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" containerName="barbican-db-sync" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.934293 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.945584 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pldnx" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.946098 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 01 16:05:07 crc kubenswrapper[4688]: I1001 16:05:07.947346 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.004988 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-74574df4f9-v4dnz"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.024937 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.034659 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-df57dbd74-8gcrt"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.039802 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.047295 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.047401 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-combined-ca-bundle\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.047453 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d13b3026-f303-454c-a686-7bc22f5211cf-logs\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.047481 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data-custom\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.047543 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6gx\" (UniqueName: \"kubernetes.io/projected/d13b3026-f303-454c-a686-7bc22f5211cf-kube-api-access-rt6gx\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.108118 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74574df4f9-v4dnz"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149811 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149850 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149892 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-combined-ca-bundle\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149911 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data-custom\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149943 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b07391ac-b11b-4b46-8e9b-847b424e7a80-logs\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149977 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-combined-ca-bundle\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.149996 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d13b3026-f303-454c-a686-7bc22f5211cf-logs\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.150017 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data-custom\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.150035 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6gx\" (UniqueName: \"kubernetes.io/projected/d13b3026-f303-454c-a686-7bc22f5211cf-kube-api-access-rt6gx\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.150061 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r784h\" (UniqueName: \"kubernetes.io/projected/b07391ac-b11b-4b46-8e9b-847b424e7a80-kube-api-access-r784h\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.165017 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d13b3026-f303-454c-a686-7bc22f5211cf-logs\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.165376 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.166744 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.168286 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.166514 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-config-data-custom\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.169265 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d13b3026-f303-454c-a686-7bc22f5211cf-combined-ca-bundle\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.169879 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.206225 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6gx\" (UniqueName: \"kubernetes.io/projected/d13b3026-f303-454c-a686-7bc22f5211cf-kube-api-access-rt6gx\") pod \"barbican-keystone-listener-df57dbd74-8gcrt\" (UID: \"d13b3026-f303-454c-a686-7bc22f5211cf\") " pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.251683 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.251739 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.251777 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r784h\" (UniqueName: \"kubernetes.io/projected/b07391ac-b11b-4b46-8e9b-847b424e7a80-kube-api-access-r784h\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.251831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2nkc\" (UniqueName: \"kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.251873 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.256632 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.256810 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.256863 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.256956 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-combined-ca-bundle\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.256985 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data-custom\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.257063 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b07391ac-b11b-4b46-8e9b-847b424e7a80-logs\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.257629 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b07391ac-b11b-4b46-8e9b-847b424e7a80-logs\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.267177 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-combined-ca-bundle\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.269073 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data-custom\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.284292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07391ac-b11b-4b46-8e9b-847b424e7a80-config-data\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.292464 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r784h\" (UniqueName: \"kubernetes.io/projected/b07391ac-b11b-4b46-8e9b-847b424e7a80-kube-api-access-r784h\") pod \"barbican-worker-74574df4f9-v4dnz\" (UID: \"b07391ac-b11b-4b46-8e9b-847b424e7a80\") " pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.350035 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.361492 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.362882 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.363140 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2nkc\" (UniqueName: \"kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.363327 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.363449 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.363658 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.365079 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.369008 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.369701 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.370397 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.371066 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.383303 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.385087 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.409627 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74574df4f9-v4dnz" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.413469 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.418058 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2nkc\" (UniqueName: \"kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc\") pod \"dnsmasq-dns-848cf88cfc-dxhhs\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.461779 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.475184 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.475327 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.475363 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.475391 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.475447 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfrc4\" (UniqueName: \"kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.526828 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.538419 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.595088 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.603778 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfrc4\" (UniqueName: \"kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.609382 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.609618 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.609664 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.609703 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.612367 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.634236 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.638397 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.641205 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.644558 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfrc4\" (UniqueName: \"kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4\") pod \"barbican-api-596f7f8496-72lx6\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.704059 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.726840 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.824839 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.824902 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.825006 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.825027 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgsvx\" (UniqueName: \"kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.825060 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.825090 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts\") pod \"81b3c008-7340-436e-873b-aa0d41588c6c\" (UID: \"81b3c008-7340-436e-873b-aa0d41588c6c\") " Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.826081 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.837399 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts" (OuterVolumeSpecName: "scripts") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.844397 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.846407 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx" (OuterVolumeSpecName: "kube-api-access-zgsvx") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "kube-api-access-zgsvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.871042 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.874859 4688 generic.go:334] "Generic (PLEG): container finished" podID="42fa9857-02f6-4a45-9761-93083ddf5104" containerID="77433936d791e20959e176cedf7427bd592cbdef3cc6723d0c7b18814b89c22d" exitCode=2 Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.874945 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerDied","Data":"77433936d791e20959e176cedf7427bd592cbdef3cc6723d0c7b18814b89c22d"} Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.896250 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqw9p" event={"ID":"81b3c008-7340-436e-873b-aa0d41588c6c","Type":"ContainerDied","Data":"607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d"} Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.896318 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="607f56c726dfcd0ef6ab4f9ec7c59c49a4ca18cef1bb30cf473f8fe963c6a82d" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.896394 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqw9p" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.927502 4688 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81b3c008-7340-436e-873b-aa0d41588c6c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.927766 4688 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.927780 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgsvx\" (UniqueName: \"kubernetes.io/projected/81b3c008-7340-436e-873b-aa0d41588c6c-kube-api-access-zgsvx\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.927793 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.927804 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:08 crc kubenswrapper[4688]: I1001 16:05:08.946862 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data" (OuterVolumeSpecName: "config-data") pod "81b3c008-7340-436e-873b-aa0d41588c6c" (UID: "81b3c008-7340-436e-873b-aa0d41588c6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.032923 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81b3c008-7340-436e-873b-aa0d41588c6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:09 crc kubenswrapper[4688]: W1001 16:05:09.078122 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd13b3026_f303_454c_a686_7bc22f5211cf.slice/crio-37f91d33c430d38b5d1f7f95511a3c0079d97b0a98618064a8978d9057196765 WatchSource:0}: Error finding container 37f91d33c430d38b5d1f7f95511a3c0079d97b0a98618064a8978d9057196765: Status 404 returned error can't find the container with id 37f91d33c430d38b5d1f7f95511a3c0079d97b0a98618064a8978d9057196765 Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.083723 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:09 crc kubenswrapper[4688]: E1001 16:05:09.084080 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" containerName="cinder-db-sync" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.084095 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" containerName="cinder-db-sync" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.084362 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" containerName="cinder-db-sync" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.085505 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.095439 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.097419 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-df57dbd74-8gcrt"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.117924 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.178155 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.210691 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.212049 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.237452 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.247722 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.247897 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbmlp\" (UniqueName: \"kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.248008 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.248054 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.248080 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.259341 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350502 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350584 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350643 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350708 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350744 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350800 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbmlp\" (UniqueName: \"kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350888 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350951 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.350986 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjv75\" (UniqueName: \"kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.351040 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.351079 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.351153 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.356316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.364139 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.364808 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.366107 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.369679 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.447032 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbmlp\" (UniqueName: \"kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp\") pod \"cinder-scheduler-0\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.449608 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.479985 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.480454 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.482150 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.482200 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.482330 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.482391 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjv75\" (UniqueName: \"kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.484902 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.485483 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.485664 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.486342 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.486756 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.499288 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.505452 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjv75\" (UniqueName: \"kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75\") pod \"dnsmasq-dns-6578955fd5-q5p5z\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.568210 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.598154 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.598253 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.600306 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.628237 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74574df4f9-v4dnz"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.646728 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:09 crc kubenswrapper[4688]: W1001 16:05:09.659678 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb07391ac_b11b_4b46_8e9b_847b424e7a80.slice/crio-66066e87837f8d64cb9b2b521f95750d6c9bf5ac850a229ad3df8f514ee730d9 WatchSource:0}: Error finding container 66066e87837f8d64cb9b2b521f95750d6c9bf5ac850a229ad3df8f514ee730d9: Status 404 returned error can't find the container with id 66066e87837f8d64cb9b2b521f95750d6c9bf5ac850a229ad3df8f514ee730d9 Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.703779 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9qjp\" (UniqueName: \"kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.703842 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.703873 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.703914 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.703963 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.704002 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.704031 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: W1001 16:05:09.721200 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod031f4890_4dc2_454b_9d92_5665d1f095fd.slice/crio-14a76f16a3fa69ebed3671af6682d78b8dcb004dce296b00988eb74b7686b38c WatchSource:0}: Error finding container 14a76f16a3fa69ebed3671af6682d78b8dcb004dce296b00988eb74b7686b38c: Status 404 returned error can't find the container with id 14a76f16a3fa69ebed3671af6682d78b8dcb004dce296b00988eb74b7686b38c Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.807910 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9qjp\" (UniqueName: \"kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.807955 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.807975 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.808001 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.808032 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.808057 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.808079 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.810801 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.812145 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.822031 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.829267 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.833188 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.837687 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.843007 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9qjp\" (UniqueName: \"kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp\") pod \"cinder-api-0\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " pod="openstack/cinder-api-0" Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.876614 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.925245 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" event={"ID":"d13b3026-f303-454c-a686-7bc22f5211cf","Type":"ContainerStarted","Data":"37f91d33c430d38b5d1f7f95511a3c0079d97b0a98618064a8978d9057196765"} Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.928361 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74574df4f9-v4dnz" event={"ID":"b07391ac-b11b-4b46-8e9b-847b424e7a80","Type":"ContainerStarted","Data":"66066e87837f8d64cb9b2b521f95750d6c9bf5ac850a229ad3df8f514ee730d9"} Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.958789 4688 generic.go:334] "Generic (PLEG): container finished" podID="42fa9857-02f6-4a45-9761-93083ddf5104" containerID="d4a9a43158e938fbb69326381e5b182bdb47e0f1e8f611b017e1bee9b41d1c0c" exitCode=0 Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.958899 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerDied","Data":"d4a9a43158e938fbb69326381e5b182bdb47e0f1e8f611b017e1bee9b41d1c0c"} Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.973798 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerStarted","Data":"181dbd26752f810c1b76e30009abb6c29902c150b9febe1287ffadde9ae526d5"} Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.981801 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" event={"ID":"031f4890-4dc2-454b-9d92-5665d1f095fd","Type":"ContainerStarted","Data":"14a76f16a3fa69ebed3671af6682d78b8dcb004dce296b00988eb74b7686b38c"} Oct 01 16:05:09 crc kubenswrapper[4688]: I1001 16:05:09.993042 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:10 crc kubenswrapper[4688]: I1001 16:05:10.375911 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:10 crc kubenswrapper[4688]: I1001 16:05:10.525029 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:05:10 crc kubenswrapper[4688]: W1001 16:05:10.610700 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ff4e6f3_7434_463c_9701_417d73f0a4bb.slice/crio-eeacaf2a627648e48914abefd726c9439a746815f713919c5089dc5328e3a30a WatchSource:0}: Error finding container eeacaf2a627648e48914abefd726c9439a746815f713919c5089dc5328e3a30a: Status 404 returned error can't find the container with id eeacaf2a627648e48914abefd726c9439a746815f713919c5089dc5328e3a30a Oct 01 16:05:10 crc kubenswrapper[4688]: I1001 16:05:10.791373 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:10 crc kubenswrapper[4688]: W1001 16:05:10.813874 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd743c9fb_e5c2_4b4b_bad5_99be991c33f9.slice/crio-fd20738176fdf949ba697ded1fbc1a2d2857d8a63e08ca08671d1c889e818ddc WatchSource:0}: Error finding container fd20738176fdf949ba697ded1fbc1a2d2857d8a63e08ca08671d1c889e818ddc: Status 404 returned error can't find the container with id fd20738176fdf949ba697ded1fbc1a2d2857d8a63e08ca08671d1c889e818ddc Oct 01 16:05:10 crc kubenswrapper[4688]: I1001 16:05:10.994367 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerStarted","Data":"fd20738176fdf949ba697ded1fbc1a2d2857d8a63e08ca08671d1c889e818ddc"} Oct 01 16:05:10 crc kubenswrapper[4688]: I1001 16:05:10.996971 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerStarted","Data":"8ed7d67cd8eac35d84ff5ea44fda06b5d178255ef0695344b65a418c6e40396c"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.006912 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerStarted","Data":"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.006952 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerStarted","Data":"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.007962 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.007994 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.013729 4688 generic.go:334] "Generic (PLEG): container finished" podID="031f4890-4dc2-454b-9d92-5665d1f095fd" containerID="7e9bd0e50d64977922276e0f0ca7a1d79c0354fa9fe1c55049c8a959d4e44e16" exitCode=0 Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.013795 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" event={"ID":"031f4890-4dc2-454b-9d92-5665d1f095fd","Type":"ContainerDied","Data":"7e9bd0e50d64977922276e0f0ca7a1d79c0354fa9fe1c55049c8a959d4e44e16"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.018369 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerStarted","Data":"ae5a5f263d62c9c6df8c85192db68e99dc40a7f87a8d652e58685339c0e4fb66"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.018406 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerStarted","Data":"eeacaf2a627648e48914abefd726c9439a746815f713919c5089dc5328e3a30a"} Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.029972 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-596f7f8496-72lx6" podStartSLOduration=3.029953173 podStartE2EDuration="3.029953173s" podCreationTimestamp="2025-10-01 16:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:11.026699332 +0000 UTC m=+1100.377339304" watchObservedRunningTime="2025-10-01 16:05:11.029953173 +0000 UTC m=+1100.380593135" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.697681 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874588 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874678 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874753 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874793 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874824 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.874850 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2nkc\" (UniqueName: \"kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc\") pod \"031f4890-4dc2-454b-9d92-5665d1f095fd\" (UID: \"031f4890-4dc2-454b-9d92-5665d1f095fd\") " Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.890764 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc" (OuterVolumeSpecName: "kube-api-access-x2nkc") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "kube-api-access-x2nkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.936408 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.977032 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.977299 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2nkc\" (UniqueName: \"kubernetes.io/projected/031f4890-4dc2-454b-9d92-5665d1f095fd-kube-api-access-x2nkc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.977313 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.977324 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.977702 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config" (OuterVolumeSpecName: "config") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:11 crc kubenswrapper[4688]: I1001 16:05:11.991153 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.010138 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "031f4890-4dc2-454b-9d92-5665d1f095fd" (UID: "031f4890-4dc2-454b-9d92-5665d1f095fd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.060510 4688 generic.go:334] "Generic (PLEG): container finished" podID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerID="ae5a5f263d62c9c6df8c85192db68e99dc40a7f87a8d652e58685339c0e4fb66" exitCode=0 Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.060628 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerDied","Data":"ae5a5f263d62c9c6df8c85192db68e99dc40a7f87a8d652e58685339c0e4fb66"} Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.079811 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.079838 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.079848 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/031f4890-4dc2-454b-9d92-5665d1f095fd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.081047 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" event={"ID":"031f4890-4dc2-454b-9d92-5665d1f095fd","Type":"ContainerDied","Data":"14a76f16a3fa69ebed3671af6682d78b8dcb004dce296b00988eb74b7686b38c"} Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.081085 4688 scope.go:117] "RemoveContainer" containerID="7e9bd0e50d64977922276e0f0ca7a1d79c0354fa9fe1c55049c8a959d4e44e16" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.081049 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-dxhhs" Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.192360 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.230778 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-dxhhs"] Oct 01 16:05:12 crc kubenswrapper[4688]: I1001 16:05:12.492416 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:13 crc kubenswrapper[4688]: I1001 16:05:13.092025 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerStarted","Data":"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b"} Oct 01 16:05:13 crc kubenswrapper[4688]: I1001 16:05:13.094508 4688 generic.go:334] "Generic (PLEG): container finished" podID="42fa9857-02f6-4a45-9761-93083ddf5104" containerID="2b7f7e054af4f411521fe37a0501664525292ea06e93ed75734516903533122c" exitCode=0 Oct 01 16:05:13 crc kubenswrapper[4688]: I1001 16:05:13.094580 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerDied","Data":"2b7f7e054af4f411521fe37a0501664525292ea06e93ed75734516903533122c"} Oct 01 16:05:13 crc kubenswrapper[4688]: I1001 16:05:13.414135 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="031f4890-4dc2-454b-9d92-5665d1f095fd" path="/var/lib/kubelet/pods/031f4890-4dc2-454b-9d92-5665d1f095fd/volumes" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.117869 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerStarted","Data":"b5e01ee8496aba3e1f6df99b638140f61b7dcae71b6229a1d9f47f172dd56e97"} Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.121248 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.137615 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerStarted","Data":"7868c87faa6576e94dd7ff0763a656e3bcba7a3aa394f39364713a4fe13bad78"} Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.150899 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42fa9857-02f6-4a45-9761-93083ddf5104","Type":"ContainerDied","Data":"bbdd091dad51544a67eb44e3ba2a3aba08ab1fcdd81e9a58edf44b4ab68f597a"} Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.151149 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbdd091dad51544a67eb44e3ba2a3aba08ab1fcdd81e9a58edf44b4ab68f597a" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.157156 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" podStartSLOduration=5.157140317 podStartE2EDuration="5.157140317s" podCreationTimestamp="2025-10-01 16:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:14.151579672 +0000 UTC m=+1103.502219634" watchObservedRunningTime="2025-10-01 16:05:14.157140317 +0000 UTC m=+1103.507780279" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.257566 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.439377 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.439433 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.439952 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440015 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440316 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440352 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440410 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5wbc\" (UniqueName: \"kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440429 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd\") pod \"42fa9857-02f6-4a45-9761-93083ddf5104\" (UID: \"42fa9857-02f6-4a45-9761-93083ddf5104\") " Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.440766 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.441084 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.458806 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc" (OuterVolumeSpecName: "kube-api-access-d5wbc") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "kube-api-access-d5wbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.467330 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts" (OuterVolumeSpecName: "scripts") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.542950 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5wbc\" (UniqueName: \"kubernetes.io/projected/42fa9857-02f6-4a45-9761-93083ddf5104-kube-api-access-d5wbc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.543285 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42fa9857-02f6-4a45-9761-93083ddf5104-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.543296 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.592753 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.627633 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data" (OuterVolumeSpecName: "config-data") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.647500 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.647540 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.673643 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42fa9857-02f6-4a45-9761-93083ddf5104" (UID: "42fa9857-02f6-4a45-9761-93083ddf5104"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.748674 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fa9857-02f6-4a45-9761-93083ddf5104-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:14 crc kubenswrapper[4688]: I1001 16:05:14.964703 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.163304 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" event={"ID":"d13b3026-f303-454c-a686-7bc22f5211cf","Type":"ContainerStarted","Data":"6e5b2036344ed72fed52166f339f1d5e106bc616862ca127aaaa1937d0b085fe"} Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.163622 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" event={"ID":"d13b3026-f303-454c-a686-7bc22f5211cf","Type":"ContainerStarted","Data":"5207d095a821f46de1f667f085b0c1036e028cf551608aa6c9d8e2ea0eac50ac"} Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.182114 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.182056 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74574df4f9-v4dnz" event={"ID":"b07391ac-b11b-4b46-8e9b-847b424e7a80","Type":"ContainerStarted","Data":"977266046db8b3791cac0604b7b6788b68c29c6957cbb7d70ec6ccaba9acc337"} Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.182656 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74574df4f9-v4dnz" event={"ID":"b07391ac-b11b-4b46-8e9b-847b424e7a80","Type":"ContainerStarted","Data":"cb1db77adaa9b37766c89fdd6505addd407f704349bdcde91eaf4b8b6e919056"} Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.209037 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-df57dbd74-8gcrt" podStartSLOduration=3.09765697 podStartE2EDuration="8.208991555s" podCreationTimestamp="2025-10-01 16:05:07 +0000 UTC" firstStartedPulling="2025-10-01 16:05:09.092000207 +0000 UTC m=+1098.442640169" lastFinishedPulling="2025-10-01 16:05:14.203334792 +0000 UTC m=+1103.553974754" observedRunningTime="2025-10-01 16:05:15.19298006 +0000 UTC m=+1104.543620032" watchObservedRunningTime="2025-10-01 16:05:15.208991555 +0000 UTC m=+1104.559631527" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.245213 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-74574df4f9-v4dnz" podStartSLOduration=3.8476306 podStartE2EDuration="8.245194862s" podCreationTimestamp="2025-10-01 16:05:07 +0000 UTC" firstStartedPulling="2025-10-01 16:05:09.680078794 +0000 UTC m=+1099.030718756" lastFinishedPulling="2025-10-01 16:05:14.077643056 +0000 UTC m=+1103.428283018" observedRunningTime="2025-10-01 16:05:15.242281481 +0000 UTC m=+1104.592921443" watchObservedRunningTime="2025-10-01 16:05:15.245194862 +0000 UTC m=+1104.595834824" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.304403 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.332568 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.355577 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:15 crc kubenswrapper[4688]: E1001 16:05:15.356318 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="sg-core" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356336 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="sg-core" Oct 01 16:05:15 crc kubenswrapper[4688]: E1001 16:05:15.356346 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="proxy-httpd" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356353 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="proxy-httpd" Oct 01 16:05:15 crc kubenswrapper[4688]: E1001 16:05:15.356365 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="ceilometer-notification-agent" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356371 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="ceilometer-notification-agent" Oct 01 16:05:15 crc kubenswrapper[4688]: E1001 16:05:15.356398 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="031f4890-4dc2-454b-9d92-5665d1f095fd" containerName="init" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356404 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="031f4890-4dc2-454b-9d92-5665d1f095fd" containerName="init" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356571 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="proxy-httpd" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356593 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="031f4890-4dc2-454b-9d92-5665d1f095fd" containerName="init" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356604 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="ceilometer-notification-agent" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.356627 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" containerName="sg-core" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.358172 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.362515 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.365055 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.374022 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.408938 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42fa9857-02f6-4a45-9761-93083ddf5104" path="/var/lib/kubelet/pods/42fa9857-02f6-4a45-9761-93083ddf5104/volumes" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471661 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471719 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471740 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471767 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471788 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9267\" (UniqueName: \"kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471806 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.471850 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573612 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573671 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573692 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573717 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573746 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9267\" (UniqueName: \"kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573765 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.573809 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.576976 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.578240 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.596773 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.597877 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.608146 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.614934 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.628076 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9267\" (UniqueName: \"kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267\") pod \"ceilometer-0\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.843197 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.908835 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8668d66c66-j8wfd"] Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.935117 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8668d66c66-j8wfd"] Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.935208 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.940127 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.940338 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 01 16:05:15 crc kubenswrapper[4688]: I1001 16:05:15.944698 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091383 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts\") pod \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091504 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key\") pod \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091549 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhdt4\" (UniqueName: \"kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4\") pod \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091745 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs\") pod \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091760 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data\") pod \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\" (UID: \"c5351885-4b4f-4e52-9472-9e5f0bf69a2f\") " Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091965 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data-custom\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.091985 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-combined-ca-bundle\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092076 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-internal-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092102 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6de36fb1-5b60-4600-ae35-bd31fdec91ce-logs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092125 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-public-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092159 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092177 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtb8v\" (UniqueName: \"kubernetes.io/projected/6de36fb1-5b60-4600-ae35-bd31fdec91ce-kube-api-access-rtb8v\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.092578 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs" (OuterVolumeSpecName: "logs") pod "c5351885-4b4f-4e52-9472-9e5f0bf69a2f" (UID: "c5351885-4b4f-4e52-9472-9e5f0bf69a2f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.098499 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c5351885-4b4f-4e52-9472-9e5f0bf69a2f" (UID: "c5351885-4b4f-4e52-9472-9e5f0bf69a2f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.114190 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4" (OuterVolumeSpecName: "kube-api-access-qhdt4") pod "c5351885-4b4f-4e52-9472-9e5f0bf69a2f" (UID: "c5351885-4b4f-4e52-9472-9e5f0bf69a2f"). InnerVolumeSpecName "kube-api-access-qhdt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.136638 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts" (OuterVolumeSpecName: "scripts") pod "c5351885-4b4f-4e52-9472-9e5f0bf69a2f" (UID: "c5351885-4b4f-4e52-9472-9e5f0bf69a2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.144717 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data" (OuterVolumeSpecName: "config-data") pod "c5351885-4b4f-4e52-9472-9e5f0bf69a2f" (UID: "c5351885-4b4f-4e52-9472-9e5f0bf69a2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.195965 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data-custom\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196019 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-combined-ca-bundle\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196106 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-internal-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196138 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6de36fb1-5b60-4600-ae35-bd31fdec91ce-logs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196169 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-public-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196209 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196231 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtb8v\" (UniqueName: \"kubernetes.io/projected/6de36fb1-5b60-4600-ae35-bd31fdec91ce-kube-api-access-rtb8v\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196316 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196329 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196339 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196351 4688 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.196361 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhdt4\" (UniqueName: \"kubernetes.io/projected/c5351885-4b4f-4e52-9472-9e5f0bf69a2f-kube-api-access-qhdt4\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.199536 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6de36fb1-5b60-4600-ae35-bd31fdec91ce-logs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.210989 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-public-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.228124 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data-custom\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.247835 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-config-data\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.248365 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-combined-ca-bundle\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.248423 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerStarted","Data":"9c560aa6b1cbc763d6cd9b5e20fa2f03e5510eb677dba9d598a31df3f791a9ca"} Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.248667 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6de36fb1-5b60-4600-ae35-bd31fdec91ce-internal-tls-certs\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.251083 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerStarted","Data":"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80"} Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.251259 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.251258 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api-log" containerID="cri-o://96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" gracePeriod=30 Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.251388 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api" containerID="cri-o://709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" gracePeriod=30 Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.252741 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtb8v\" (UniqueName: \"kubernetes.io/projected/6de36fb1-5b60-4600-ae35-bd31fdec91ce-kube-api-access-rtb8v\") pod \"barbican-api-8668d66c66-j8wfd\" (UID: \"6de36fb1-5b60-4600-ae35-bd31fdec91ce\") " pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.281520 4688 generic.go:334] "Generic (PLEG): container finished" podID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerID="047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" exitCode=137 Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.281566 4688 generic.go:334] "Generic (PLEG): container finished" podID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerID="93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" exitCode=137 Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.282307 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5889fcf659-ptwxd" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.283092 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.155021901 podStartE2EDuration="7.2830762s" podCreationTimestamp="2025-10-01 16:05:09 +0000 UTC" firstStartedPulling="2025-10-01 16:05:10.429685665 +0000 UTC m=+1099.780325627" lastFinishedPulling="2025-10-01 16:05:11.557739964 +0000 UTC m=+1100.908379926" observedRunningTime="2025-10-01 16:05:16.280246472 +0000 UTC m=+1105.630886434" watchObservedRunningTime="2025-10-01 16:05:16.2830762 +0000 UTC m=+1105.633716162" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.288210 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerDied","Data":"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb"} Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.288245 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerDied","Data":"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8"} Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.288255 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5889fcf659-ptwxd" event={"ID":"c5351885-4b4f-4e52-9472-9e5f0bf69a2f","Type":"ContainerDied","Data":"414ed783e209ce593cad2d7110a77218656be1b3c6e8f544432424a7fb2443dc"} Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.288269 4688 scope.go:117] "RemoveContainer" containerID="047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.318956 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.3189347080000005 podStartE2EDuration="7.318934708s" podCreationTimestamp="2025-10-01 16:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:16.304410544 +0000 UTC m=+1105.655050526" watchObservedRunningTime="2025-10-01 16:05:16.318934708 +0000 UTC m=+1105.669574670" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.375075 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.418492 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5889fcf659-ptwxd"] Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.460010 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.561300 4688 scope.go:117] "RemoveContainer" containerID="93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.562043 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:16 crc kubenswrapper[4688]: W1001 16:05:16.611212 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13321c8d_8fe9_4036_9061_0f5466c84b6c.slice/crio-a3c8d852dee5a9d4d93242e3a849ecc6ede8ab743e0b0c443fa5e338343b53d2 WatchSource:0}: Error finding container a3c8d852dee5a9d4d93242e3a849ecc6ede8ab743e0b0c443fa5e338343b53d2: Status 404 returned error can't find the container with id a3c8d852dee5a9d4d93242e3a849ecc6ede8ab743e0b0c443fa5e338343b53d2 Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.634208 4688 scope.go:117] "RemoveContainer" containerID="047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" Oct 01 16:05:16 crc kubenswrapper[4688]: E1001 16:05:16.634966 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb\": container with ID starting with 047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb not found: ID does not exist" containerID="047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635007 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb"} err="failed to get container status \"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb\": rpc error: code = NotFound desc = could not find container \"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb\": container with ID starting with 047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb not found: ID does not exist" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635034 4688 scope.go:117] "RemoveContainer" containerID="93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" Oct 01 16:05:16 crc kubenswrapper[4688]: E1001 16:05:16.635538 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8\": container with ID starting with 93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8 not found: ID does not exist" containerID="93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635558 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8"} err="failed to get container status \"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8\": rpc error: code = NotFound desc = could not find container \"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8\": container with ID starting with 93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8 not found: ID does not exist" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635572 4688 scope.go:117] "RemoveContainer" containerID="047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635816 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb"} err="failed to get container status \"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb\": rpc error: code = NotFound desc = could not find container \"047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb\": container with ID starting with 047aa88c09008efd9c51d79d32665b3a292d6be912e017248d58a4ba3f3637bb not found: ID does not exist" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.635832 4688 scope.go:117] "RemoveContainer" containerID="93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8" Oct 01 16:05:16 crc kubenswrapper[4688]: I1001 16:05:16.636090 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8"} err="failed to get container status \"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8\": rpc error: code = NotFound desc = could not find container \"93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8\": container with ID starting with 93e2f9a3527158aed14e6779d5ff099107172eb33f33ba94d007e294459090b8 not found: ID does not exist" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.119897 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8668d66c66-j8wfd"] Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.242042 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.315939 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8668d66c66-j8wfd" event={"ID":"6de36fb1-5b60-4600-ae35-bd31fdec91ce","Type":"ContainerStarted","Data":"7991b8615491bfff27ab5dcdee5cc67e498cafdfd1df9a3de58de96421ef2843"} Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340105 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340219 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340254 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340281 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340319 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340393 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9qjp\" (UniqueName: \"kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.340496 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id\") pod \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\" (UID: \"d743c9fb-e5c2-4b4b-bad5-99be991c33f9\") " Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.341142 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.341998 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs" (OuterVolumeSpecName: "logs") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345634 4688 generic.go:334] "Generic (PLEG): container finished" podID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerID="709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" exitCode=0 Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345662 4688 generic.go:334] "Generic (PLEG): container finished" podID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerID="96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" exitCode=143 Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345725 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerDied","Data":"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80"} Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345751 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerDied","Data":"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b"} Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345760 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d743c9fb-e5c2-4b4b-bad5-99be991c33f9","Type":"ContainerDied","Data":"fd20738176fdf949ba697ded1fbc1a2d2857d8a63e08ca08671d1c889e818ddc"} Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345795 4688 scope.go:117] "RemoveContainer" containerID="709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.345935 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.349584 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.359624 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerStarted","Data":"a3c8d852dee5a9d4d93242e3a849ecc6ede8ab743e0b0c443fa5e338343b53d2"} Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.363704 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp" (OuterVolumeSpecName: "kube-api-access-z9qjp") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "kube-api-access-z9qjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.364775 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts" (OuterVolumeSpecName: "scripts") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.396751 4688 scope.go:117] "RemoveContainer" containerID="96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.406586 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" path="/var/lib/kubelet/pods/c5351885-4b4f-4e52-9472-9e5f0bf69a2f/volumes" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.417682 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.427816 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data" (OuterVolumeSpecName: "config-data") pod "d743c9fb-e5c2-4b4b-bad5-99be991c33f9" (UID: "d743c9fb-e5c2-4b4b-bad5-99be991c33f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.435679 4688 scope.go:117] "RemoveContainer" containerID="709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.436295 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80\": container with ID starting with 709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80 not found: ID does not exist" containerID="709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.436325 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80"} err="failed to get container status \"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80\": rpc error: code = NotFound desc = could not find container \"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80\": container with ID starting with 709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80 not found: ID does not exist" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.436343 4688 scope.go:117] "RemoveContainer" containerID="96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.436899 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b\": container with ID starting with 96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b not found: ID does not exist" containerID="96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.436923 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b"} err="failed to get container status \"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b\": rpc error: code = NotFound desc = could not find container \"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b\": container with ID starting with 96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b not found: ID does not exist" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.436936 4688 scope.go:117] "RemoveContainer" containerID="709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.437304 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80"} err="failed to get container status \"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80\": rpc error: code = NotFound desc = could not find container \"709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80\": container with ID starting with 709dcc88006e518bc728d4582ad86086ff28463651a8ee19ce316f79c1ab4c80 not found: ID does not exist" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.437326 4688 scope.go:117] "RemoveContainer" containerID="96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.437547 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b"} err="failed to get container status \"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b\": rpc error: code = NotFound desc = could not find container \"96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b\": container with ID starting with 96696dd1edab3f963489942924f79d242298ec322fbc25f969965fb1eead022b not found: ID does not exist" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447621 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447643 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9qjp\" (UniqueName: \"kubernetes.io/projected/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-kube-api-access-z9qjp\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447653 4688 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447661 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447669 4688 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447677 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.447686 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d743c9fb-e5c2-4b4b-bad5-99be991c33f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.695045 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.730977 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.742594 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.742994 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743008 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon" Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.743021 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743027 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api" Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.743038 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon-log" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743044 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon-log" Oct 01 16:05:17 crc kubenswrapper[4688]: E1001 16:05:17.743064 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api-log" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743069 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api-log" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743262 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api-log" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743286 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" containerName="cinder-api" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743295 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.743304 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5351885-4b4f-4e52-9472-9e5f0bf69a2f" containerName="horizon-log" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.744309 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.748376 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.749928 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.750028 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.750162 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753091 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753214 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753317 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data-custom\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753391 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk7rq\" (UniqueName: \"kubernetes.io/projected/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-kube-api-access-gk7rq\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753487 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-scripts\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753583 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-logs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753664 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753733 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.753793 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855031 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data-custom\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855290 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk7rq\" (UniqueName: \"kubernetes.io/projected/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-kube-api-access-gk7rq\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855397 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-scripts\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855466 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-logs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855560 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855663 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855737 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855865 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.855946 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.856122 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-logs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.856249 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.862293 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.864498 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.865841 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.867100 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-config-data-custom\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.878125 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-scripts\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.878207 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk7rq\" (UniqueName: \"kubernetes.io/projected/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-kube-api-access-gk7rq\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.878342 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12a98d94-2c07-4fa8-b69b-c24a11fc6cac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"12a98d94-2c07-4fa8-b69b-c24a11fc6cac\") " pod="openstack/cinder-api-0" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.886703 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c946d4bf9-l8zds" Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.967561 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.967757 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-f8467544d-9fz8w" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-api" containerID="cri-o://73abce7839d4c71f72b092bcafbba0a7ec6f26e945f08b62e9c618f93d3b4171" gracePeriod=30 Oct 01 16:05:17 crc kubenswrapper[4688]: I1001 16:05:17.967897 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-f8467544d-9fz8w" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-httpd" containerID="cri-o://c00340087753728353eab21fd7726560d201e35cc7849e4943ecbf0cdf0504d0" gracePeriod=30 Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.072021 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.444755 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8668d66c66-j8wfd" event={"ID":"6de36fb1-5b60-4600-ae35-bd31fdec91ce","Type":"ContainerStarted","Data":"d436fcf9abad007fc5fd7cbb662fd30bade2f1d7a1ddb3cc840152903ba66ccd"} Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.445073 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8668d66c66-j8wfd" event={"ID":"6de36fb1-5b60-4600-ae35-bd31fdec91ce","Type":"ContainerStarted","Data":"7c596ce6f2f8b4c77bf63df1f5e9928ab49ede866f84f0565254b8055ea809f9"} Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.446210 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.446241 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.457016 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerStarted","Data":"2a3d18b965114473d044a618e486a02e57613031dd297e6a780b4204a5af48ed"} Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.487641 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8668d66c66-j8wfd" podStartSLOduration=3.487618221 podStartE2EDuration="3.487618221s" podCreationTimestamp="2025-10-01 16:05:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:18.48543634 +0000 UTC m=+1107.836076302" watchObservedRunningTime="2025-10-01 16:05:18.487618221 +0000 UTC m=+1107.838258183" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.538168 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.538238 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.538931 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2"} pod="openstack/horizon-654b94b7dd-g2k9q" containerMessage="Container horizon failed startup probe, will be restarted" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.538974 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" containerID="cri-o://db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2" gracePeriod=30 Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.588753 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.588817 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.589471 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"674bcb3920fc3555ff6521c9241431dd30559cff036703357069642077847a95"} pod="openstack/horizon-555656bf5b-xlntq" containerMessage="Container horizon failed startup probe, will be restarted" Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.589501 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" containerID="cri-o://674bcb3920fc3555ff6521c9241431dd30559cff036703357069642077847a95" gracePeriod=30 Oct 01 16:05:18 crc kubenswrapper[4688]: I1001 16:05:18.800094 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.403039 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d743c9fb-e5c2-4b4b-bad5-99be991c33f9" path="/var/lib/kubelet/pods/d743c9fb-e5c2-4b4b-bad5-99be991c33f9/volumes" Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.455229 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.491874 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerStarted","Data":"37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7"} Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.494370 4688 generic.go:334] "Generic (PLEG): container finished" podID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerID="c00340087753728353eab21fd7726560d201e35cc7849e4943ecbf0cdf0504d0" exitCode=0 Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.494438 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerDied","Data":"c00340087753728353eab21fd7726560d201e35cc7849e4943ecbf0cdf0504d0"} Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.503587 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12a98d94-2c07-4fa8-b69b-c24a11fc6cac","Type":"ContainerStarted","Data":"39b8f0a4ef198621d503143d6fff0eab180ca1a4bbd3f1f6620ab3fb168eebf5"} Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.570498 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.646059 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.646332 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="dnsmasq-dns" containerID="cri-o://534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a" gracePeriod=10 Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.916774 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Oct 01 16:05:19 crc kubenswrapper[4688]: I1001 16:05:19.960439 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.064558 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.474023 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.558759 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerStarted","Data":"082c75db6aab9a4dae6032996e3712882be3e10e3b4573c57ce5d1421fee3265"} Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.585255 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12a98d94-2c07-4fa8-b69b-c24a11fc6cac","Type":"ContainerStarted","Data":"79634f95a1ae3a48a5a450af30a1dc8ecc2965de449d16ede462b35a101fee15"} Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.627758 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lfh9\" (UniqueName: \"kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.627828 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.627868 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.627899 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.627959 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.628002 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb\") pod \"ef2b662d-30ed-497e-84e2-a4cdde464337\" (UID: \"ef2b662d-30ed-497e-84e2-a4cdde464337\") " Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.633326 4688 generic.go:334] "Generic (PLEG): container finished" podID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerID="534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a" exitCode=0 Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.639496 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.641825 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9" (OuterVolumeSpecName: "kube-api-access-2lfh9") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "kube-api-access-2lfh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.641962 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" event={"ID":"ef2b662d-30ed-497e-84e2-a4cdde464337","Type":"ContainerDied","Data":"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a"} Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.642054 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-k7bhx" event={"ID":"ef2b662d-30ed-497e-84e2-a4cdde464337","Type":"ContainerDied","Data":"f666241b4207d73a8c49fc648bc899f0ddd52329c1387da1b4b355a31cde9ea4"} Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.642102 4688 scope.go:117] "RemoveContainer" containerID="534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.649343 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="cinder-scheduler" containerID="cri-o://7868c87faa6576e94dd7ff0763a656e3bcba7a3aa394f39364713a4fe13bad78" gracePeriod=30 Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.649764 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="probe" containerID="cri-o://9c560aa6b1cbc763d6cd9b5e20fa2f03e5510eb677dba9d598a31df3f791a9ca" gracePeriod=30 Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.728832 4688 scope.go:117] "RemoveContainer" containerID="fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.730735 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lfh9\" (UniqueName: \"kubernetes.io/projected/ef2b662d-30ed-497e-84e2-a4cdde464337-kube-api-access-2lfh9\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.738148 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config" (OuterVolumeSpecName: "config") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.743145 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.746119 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.766175 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.810151 4688 scope.go:117] "RemoveContainer" containerID="534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a" Oct 01 16:05:20 crc kubenswrapper[4688]: E1001 16:05:20.810936 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a\": container with ID starting with 534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a not found: ID does not exist" containerID="534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.811002 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a"} err="failed to get container status \"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a\": rpc error: code = NotFound desc = could not find container \"534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a\": container with ID starting with 534e9ddcb00761a98845c619a44f51e724cc63356b1ccd57f5cad18ae15bb19a not found: ID does not exist" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.811053 4688 scope.go:117] "RemoveContainer" containerID="fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527" Oct 01 16:05:20 crc kubenswrapper[4688]: E1001 16:05:20.811371 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527\": container with ID starting with fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527 not found: ID does not exist" containerID="fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.811422 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527"} err="failed to get container status \"fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527\": rpc error: code = NotFound desc = could not find container \"fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527\": container with ID starting with fc8bbe095053dc74addd19810bff18a1416781f131967644a105ac6576759527 not found: ID does not exist" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.816822 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef2b662d-30ed-497e-84e2-a4cdde464337" (UID: "ef2b662d-30ed-497e-84e2-a4cdde464337"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.832045 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.832085 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.832101 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.832117 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.832129 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef2b662d-30ed-497e-84e2-a4cdde464337-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.977564 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:05:20 crc kubenswrapper[4688]: I1001 16:05:20.986554 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-k7bhx"] Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.391094 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" path="/var/lib/kubelet/pods/ef2b662d-30ed-497e-84e2-a4cdde464337/volumes" Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.642482 4688 generic.go:334] "Generic (PLEG): container finished" podID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerID="73abce7839d4c71f72b092bcafbba0a7ec6f26e945f08b62e9c618f93d3b4171" exitCode=0 Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.642581 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerDied","Data":"73abce7839d4c71f72b092bcafbba0a7ec6f26e945f08b62e9c618f93d3b4171"} Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.644664 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12a98d94-2c07-4fa8-b69b-c24a11fc6cac","Type":"ContainerStarted","Data":"ecb2b819a5d05760d7c3297a322b040ee2041a5a47c1549273d988b04807e2b9"} Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.644766 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.648669 4688 generic.go:334] "Generic (PLEG): container finished" podID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerID="9c560aa6b1cbc763d6cd9b5e20fa2f03e5510eb677dba9d598a31df3f791a9ca" exitCode=0 Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.648712 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerDied","Data":"9c560aa6b1cbc763d6cd9b5e20fa2f03e5510eb677dba9d598a31df3f791a9ca"} Oct 01 16:05:21 crc kubenswrapper[4688]: I1001 16:05:21.670834 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.6708201129999996 podStartE2EDuration="4.670820113s" podCreationTimestamp="2025-10-01 16:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:21.669021903 +0000 UTC m=+1111.019661875" watchObservedRunningTime="2025-10-01 16:05:21.670820113 +0000 UTC m=+1111.021460075" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.092403 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.176002 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config\") pod \"b31db0c2-3adf-4c3f-b530-87e8eece807e\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.176100 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs\") pod \"b31db0c2-3adf-4c3f-b530-87e8eece807e\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.176196 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config\") pod \"b31db0c2-3adf-4c3f-b530-87e8eece807e\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.176245 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw5cn\" (UniqueName: \"kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn\") pod \"b31db0c2-3adf-4c3f-b530-87e8eece807e\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.176266 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle\") pod \"b31db0c2-3adf-4c3f-b530-87e8eece807e\" (UID: \"b31db0c2-3adf-4c3f-b530-87e8eece807e\") " Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.184846 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn" (OuterVolumeSpecName: "kube-api-access-qw5cn") pod "b31db0c2-3adf-4c3f-b530-87e8eece807e" (UID: "b31db0c2-3adf-4c3f-b530-87e8eece807e"). InnerVolumeSpecName "kube-api-access-qw5cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.196755 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b31db0c2-3adf-4c3f-b530-87e8eece807e" (UID: "b31db0c2-3adf-4c3f-b530-87e8eece807e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.206171 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.280194 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw5cn\" (UniqueName: \"kubernetes.io/projected/b31db0c2-3adf-4c3f-b530-87e8eece807e-kube-api-access-qw5cn\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.280992 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.309727 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config" (OuterVolumeSpecName: "config") pod "b31db0c2-3adf-4c3f-b530-87e8eece807e" (UID: "b31db0c2-3adf-4c3f-b530-87e8eece807e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.382484 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.391368 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.447132 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b31db0c2-3adf-4c3f-b530-87e8eece807e" (UID: "b31db0c2-3adf-4c3f-b530-87e8eece807e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.463760 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b31db0c2-3adf-4c3f-b530-87e8eece807e" (UID: "b31db0c2-3adf-4c3f-b530-87e8eece807e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.483875 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.484071 4688 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31db0c2-3adf-4c3f-b530-87e8eece807e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.659897 4688 generic.go:334] "Generic (PLEG): container finished" podID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerID="7868c87faa6576e94dd7ff0763a656e3bcba7a3aa394f39364713a4fe13bad78" exitCode=0 Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.659982 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerDied","Data":"7868c87faa6576e94dd7ff0763a656e3bcba7a3aa394f39364713a4fe13bad78"} Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.663663 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f8467544d-9fz8w" event={"ID":"b31db0c2-3adf-4c3f-b530-87e8eece807e","Type":"ContainerDied","Data":"d4c516a2d85e1b38d9b2435307a28c8f856bc9f20bfd888e02f60137f262ee6c"} Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.663721 4688 scope.go:117] "RemoveContainer" containerID="c00340087753728353eab21fd7726560d201e35cc7849e4943ecbf0cdf0504d0" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.663729 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f8467544d-9fz8w" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.691344 4688 scope.go:117] "RemoveContainer" containerID="73abce7839d4c71f72b092bcafbba0a7ec6f26e945f08b62e9c618f93d3b4171" Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.698411 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:05:22 crc kubenswrapper[4688]: I1001 16:05:22.705392 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f8467544d-9fz8w"] Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.086980 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203127 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203278 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203349 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203414 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbmlp\" (UniqueName: \"kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203484 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.203513 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data\") pod \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\" (UID: \"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c\") " Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.209604 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.223502 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.228787 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts" (OuterVolumeSpecName: "scripts") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.236349 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp" (OuterVolumeSpecName: "kube-api-access-nbmlp") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "kube-api-access-nbmlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.311369 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.311396 4688 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.311405 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbmlp\" (UniqueName: \"kubernetes.io/projected/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-kube-api-access-nbmlp\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.311415 4688 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.359102 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.418123 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.469738 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" path="/var/lib/kubelet/pods/b31db0c2-3adf-4c3f-b530-87e8eece807e/volumes" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.506477 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data" (OuterVolumeSpecName: "config-data") pod "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" (UID: "bc97b979-5c30-4ab1-9fb3-a06c70a2c57c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.518774 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.674447 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerStarted","Data":"ce4b067d88280eee79125cdd3ff415b377d9202cab0b89ca2fd30e7cccf01ed0"} Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.675512 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.677336 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc97b979-5c30-4ab1-9fb3-a06c70a2c57c","Type":"ContainerDied","Data":"8ed7d67cd8eac35d84ff5ea44fda06b5d178255ef0695344b65a418c6e40396c"} Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.677367 4688 scope.go:117] "RemoveContainer" containerID="9c560aa6b1cbc763d6cd9b5e20fa2f03e5510eb677dba9d598a31df3f791a9ca" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.677467 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.703748 4688 scope.go:117] "RemoveContainer" containerID="7868c87faa6576e94dd7ff0763a656e3bcba7a3aa394f39364713a4fe13bad78" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.706806 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.50221675 podStartE2EDuration="8.706792854s" podCreationTimestamp="2025-10-01 16:05:15 +0000 UTC" firstStartedPulling="2025-10-01 16:05:16.615153567 +0000 UTC m=+1105.965793529" lastFinishedPulling="2025-10-01 16:05:22.819729661 +0000 UTC m=+1112.170369633" observedRunningTime="2025-10-01 16:05:23.703952535 +0000 UTC m=+1113.054592517" watchObservedRunningTime="2025-10-01 16:05:23.706792854 +0000 UTC m=+1113.057432816" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.726977 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.737059 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754080 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754415 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="dnsmasq-dns" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754431 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="dnsmasq-dns" Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754459 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-httpd" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754466 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-httpd" Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754474 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-api" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754482 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-api" Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754489 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="cinder-scheduler" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754495 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="cinder-scheduler" Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754516 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="probe" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754526 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="probe" Oct 01 16:05:23 crc kubenswrapper[4688]: E1001 16:05:23.754547 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="init" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754553 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="init" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754729 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-httpd" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754748 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="cinder-scheduler" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754765 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" containerName="probe" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754775 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b31db0c2-3adf-4c3f-b530-87e8eece807e" containerName="neutron-api" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.754783 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2b662d-30ed-497e-84e2-a4cdde464337" containerName="dnsmasq-dns" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.755629 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.760044 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.779417 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823136 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78mgp\" (UniqueName: \"kubernetes.io/projected/825e667e-354c-4cd3-bb07-d51c67981ac8-kube-api-access-78mgp\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823195 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823246 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823261 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-scripts\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823327 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825e667e-354c-4cd3-bb07-d51c67981ac8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.823389 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924656 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825e667e-354c-4cd3-bb07-d51c67981ac8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924757 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78mgp\" (UniqueName: \"kubernetes.io/projected/825e667e-354c-4cd3-bb07-d51c67981ac8-kube-api-access-78mgp\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924783 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924819 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.924834 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-scripts\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.925669 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/825e667e-354c-4cd3-bb07-d51c67981ac8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.932133 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-scripts\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.932721 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.933139 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.934916 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/825e667e-354c-4cd3-bb07-d51c67981ac8-config-data\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:23 crc kubenswrapper[4688]: I1001 16:05:23.948595 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78mgp\" (UniqueName: \"kubernetes.io/projected/825e667e-354c-4cd3-bb07-d51c67981ac8-kube-api-access-78mgp\") pod \"cinder-scheduler-0\" (UID: \"825e667e-354c-4cd3-bb07-d51c67981ac8\") " pod="openstack/cinder-scheduler-0" Oct 01 16:05:24 crc kubenswrapper[4688]: I1001 16:05:24.076147 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 16:05:24 crc kubenswrapper[4688]: W1001 16:05:24.613687 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod825e667e_354c_4cd3_bb07_d51c67981ac8.slice/crio-6041f9e1bcab7d982a5193629568bd814d519f150903cea826a4685747cd7b3c WatchSource:0}: Error finding container 6041f9e1bcab7d982a5193629568bd814d519f150903cea826a4685747cd7b3c: Status 404 returned error can't find the container with id 6041f9e1bcab7d982a5193629568bd814d519f150903cea826a4685747cd7b3c Oct 01 16:05:24 crc kubenswrapper[4688]: I1001 16:05:24.623646 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 16:05:24 crc kubenswrapper[4688]: I1001 16:05:24.692226 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825e667e-354c-4cd3-bb07-d51c67981ac8","Type":"ContainerStarted","Data":"6041f9e1bcab7d982a5193629568bd814d519f150903cea826a4685747cd7b3c"} Oct 01 16:05:25 crc kubenswrapper[4688]: I1001 16:05:25.053084 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:05:25 crc kubenswrapper[4688]: I1001 16:05:25.053140 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:05:25 crc kubenswrapper[4688]: I1001 16:05:25.399581 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc97b979-5c30-4ab1-9fb3-a06c70a2c57c" path="/var/lib/kubelet/pods/bc97b979-5c30-4ab1-9fb3-a06c70a2c57c/volumes" Oct 01 16:05:25 crc kubenswrapper[4688]: I1001 16:05:25.711279 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825e667e-354c-4cd3-bb07-d51c67981ac8","Type":"ContainerStarted","Data":"e3de5b494d0a38a702fd4fc6a6f2f3e48ccbefd866b74823c27f5da310c497e9"} Oct 01 16:05:26 crc kubenswrapper[4688]: I1001 16:05:26.720220 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"825e667e-354c-4cd3-bb07-d51c67981ac8","Type":"ContainerStarted","Data":"32d6b4c80a6750b591de14586286d6de8f6ed483e114a9e2cba6a9409633d5ff"} Oct 01 16:05:26 crc kubenswrapper[4688]: I1001 16:05:26.737864 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.737845744 podStartE2EDuration="3.737845744s" podCreationTimestamp="2025-10-01 16:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:26.736672341 +0000 UTC m=+1116.087312303" watchObservedRunningTime="2025-10-01 16:05:26.737845744 +0000 UTC m=+1116.088485696" Oct 01 16:05:27 crc kubenswrapper[4688]: I1001 16:05:27.999020 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:05:28 crc kubenswrapper[4688]: I1001 16:05:28.440277 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-76d4fdc768-gl782" Oct 01 16:05:29 crc kubenswrapper[4688]: I1001 16:05:29.077830 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 16:05:30 crc kubenswrapper[4688]: I1001 16:05:30.749951 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-587df57d87-hlpwt" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.498290 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.567767 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8668d66c66-j8wfd" podUID="6de36fb1-5b60-4600-ae35-bd31fdec91ce" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.170:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.595153 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8668d66c66-j8wfd" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.676670 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.676945 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-596f7f8496-72lx6" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api-log" containerID="cri-o://9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c" gracePeriod=30 Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.677120 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-596f7f8496-72lx6" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api" containerID="cri-o://df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784" gracePeriod=30 Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.729722 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.731037 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.739031 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zcc6n" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.739165 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.739294 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.759021 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.888743 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fccs\" (UniqueName: \"kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.888908 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.888965 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.889225 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.990580 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fccs\" (UniqueName: \"kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.990812 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.990877 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.990932 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:31 crc kubenswrapper[4688]: I1001 16:05:31.992596 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.019989 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.020387 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.048187 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fccs\" (UniqueName: \"kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs\") pod \"openstackclient\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.056403 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.080282 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="12a98d94-2c07-4fa8-b69b-c24a11fc6cac" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.099616 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.109190 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.176600 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.178017 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.194329 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.194445 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.194468 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.194507 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcplf\" (UniqueName: \"kubernetes.io/projected/a309f66f-cc9d-40e7-9acc-00881395f4c9-kube-api-access-hcplf\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.241914 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.301311 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcplf\" (UniqueName: \"kubernetes.io/projected/a309f66f-cc9d-40e7-9acc-00881395f4c9-kube-api-access-hcplf\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.301377 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.301488 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.301504 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.303192 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.309045 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.310916 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a309f66f-cc9d-40e7-9acc-00881395f4c9-openstack-config-secret\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.353326 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcplf\" (UniqueName: \"kubernetes.io/projected/a309f66f-cc9d-40e7-9acc-00881395f4c9-kube-api-access-hcplf\") pod \"openstackclient\" (UID: \"a309f66f-cc9d-40e7-9acc-00881395f4c9\") " pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: E1001 16:05:32.382118 4688 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 01 16:05:32 crc kubenswrapper[4688]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f241f15b-3967-4fc6-b25e-bbe57ae2ed77_0(532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36" Netns:"/var/run/netns/53cda106-377b-491d-95a1-fbd80b8d6f45" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36;K8S_POD_UID=f241f15b-3967-4fc6-b25e-bbe57ae2ed77" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f241f15b-3967-4fc6-b25e-bbe57ae2ed77]: expected pod UID "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" but got "a309f66f-cc9d-40e7-9acc-00881395f4c9" from Kube API Oct 01 16:05:32 crc kubenswrapper[4688]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 01 16:05:32 crc kubenswrapper[4688]: > Oct 01 16:05:32 crc kubenswrapper[4688]: E1001 16:05:32.382172 4688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 01 16:05:32 crc kubenswrapper[4688]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f241f15b-3967-4fc6-b25e-bbe57ae2ed77_0(532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36" Netns:"/var/run/netns/53cda106-377b-491d-95a1-fbd80b8d6f45" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=532fceaaef49373c13c8059cd0c4bbda665f5983cc7657e09e0f118916c50b36;K8S_POD_UID=f241f15b-3967-4fc6-b25e-bbe57ae2ed77" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f241f15b-3967-4fc6-b25e-bbe57ae2ed77]: expected pod UID "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" but got "a309f66f-cc9d-40e7-9acc-00881395f4c9" from Kube API Oct 01 16:05:32 crc kubenswrapper[4688]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 01 16:05:32 crc kubenswrapper[4688]: > pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.505400 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.785539 4688 generic.go:334] "Generic (PLEG): container finished" podID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerID="9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c" exitCode=143 Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.785609 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.786153 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerDied","Data":"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c"} Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.790327 4688 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f241f15b-3967-4fc6-b25e-bbe57ae2ed77" podUID="a309f66f-cc9d-40e7-9acc-00881395f4c9" Oct 01 16:05:32 crc kubenswrapper[4688]: I1001 16:05:32.861418 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.029723 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret\") pod \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.029885 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config\") pod \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.029952 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle\") pod \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.030017 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fccs\" (UniqueName: \"kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs\") pod \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\" (UID: \"f241f15b-3967-4fc6-b25e-bbe57ae2ed77\") " Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.030828 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" (UID: "f241f15b-3967-4fc6-b25e-bbe57ae2ed77"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.036770 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" (UID: "f241f15b-3967-4fc6-b25e-bbe57ae2ed77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.036891 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" (UID: "f241f15b-3967-4fc6-b25e-bbe57ae2ed77"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.044244 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs" (OuterVolumeSpecName: "kube-api-access-9fccs") pod "f241f15b-3967-4fc6-b25e-bbe57ae2ed77" (UID: "f241f15b-3967-4fc6-b25e-bbe57ae2ed77"). InnerVolumeSpecName "kube-api-access-9fccs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.078759 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="12a98d94-2c07-4fa8-b69b-c24a11fc6cac" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.113097 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.133005 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.133038 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fccs\" (UniqueName: \"kubernetes.io/projected/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-kube-api-access-9fccs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.133050 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.133059 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f241f15b-3967-4fc6-b25e-bbe57ae2ed77-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.399292 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f241f15b-3967-4fc6-b25e-bbe57ae2ed77" path="/var/lib/kubelet/pods/f241f15b-3967-4fc6-b25e-bbe57ae2ed77/volumes" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.802135 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.802919 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a309f66f-cc9d-40e7-9acc-00881395f4c9","Type":"ContainerStarted","Data":"e7a77d21dcacf993655079428f9d085fdf485e3add9eeca5c87fdd96e9171018"} Oct 01 16:05:33 crc kubenswrapper[4688]: I1001 16:05:33.810956 4688 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f241f15b-3967-4fc6-b25e-bbe57ae2ed77" podUID="a309f66f-cc9d-40e7-9acc-00881395f4c9" Oct 01 16:05:34 crc kubenswrapper[4688]: I1001 16:05:34.428142 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 16:05:35 crc kubenswrapper[4688]: I1001 16:05:35.854292 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.127680 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-596f7f8496-72lx6" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:39602->10.217.0.165:9311: read: connection reset by peer" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.127691 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-596f7f8496-72lx6" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:39608->10.217.0.165:9311: read: connection reset by peer" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.744580 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.841496 4688 generic.go:334] "Generic (PLEG): container finished" podID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerID="df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784" exitCode=0 Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.841578 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerDied","Data":"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784"} Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.841610 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596f7f8496-72lx6" event={"ID":"253d6dae-a9bc-4da5-be6c-0e65311d34a2","Type":"ContainerDied","Data":"181dbd26752f810c1b76e30009abb6c29902c150b9febe1287ffadde9ae526d5"} Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.841630 4688 scope.go:117] "RemoveContainer" containerID="df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.841780 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596f7f8496-72lx6" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.890722 4688 scope.go:117] "RemoveContainer" containerID="9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.913286 4688 scope.go:117] "RemoveContainer" containerID="df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784" Oct 01 16:05:36 crc kubenswrapper[4688]: E1001 16:05:36.914489 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784\": container with ID starting with df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784 not found: ID does not exist" containerID="df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.914568 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784"} err="failed to get container status \"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784\": rpc error: code = NotFound desc = could not find container \"df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784\": container with ID starting with df97614f466377a2491f1ebc41595e16416edf53e4bde3128dd2663232034784 not found: ID does not exist" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.914604 4688 scope.go:117] "RemoveContainer" containerID="9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c" Oct 01 16:05:36 crc kubenswrapper[4688]: E1001 16:05:36.915005 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c\": container with ID starting with 9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c not found: ID does not exist" containerID="9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915037 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c"} err="failed to get container status \"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c\": rpc error: code = NotFound desc = could not find container \"9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c\": container with ID starting with 9d746cf7db5675077fc9f0e573762f2ca4c8dbc840a7fce1418598f232cf4b0c not found: ID does not exist" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915170 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle\") pod \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915276 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfrc4\" (UniqueName: \"kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4\") pod \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915354 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs\") pod \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915385 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data\") pod \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.915459 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom\") pod \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\" (UID: \"253d6dae-a9bc-4da5-be6c-0e65311d34a2\") " Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.917625 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs" (OuterVolumeSpecName: "logs") pod "253d6dae-a9bc-4da5-be6c-0e65311d34a2" (UID: "253d6dae-a9bc-4da5-be6c-0e65311d34a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.921365 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4" (OuterVolumeSpecName: "kube-api-access-pfrc4") pod "253d6dae-a9bc-4da5-be6c-0e65311d34a2" (UID: "253d6dae-a9bc-4da5-be6c-0e65311d34a2"). InnerVolumeSpecName "kube-api-access-pfrc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.930704 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "253d6dae-a9bc-4da5-be6c-0e65311d34a2" (UID: "253d6dae-a9bc-4da5-be6c-0e65311d34a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.945938 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "253d6dae-a9bc-4da5-be6c-0e65311d34a2" (UID: "253d6dae-a9bc-4da5-be6c-0e65311d34a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:36 crc kubenswrapper[4688]: I1001 16:05:36.983236 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data" (OuterVolumeSpecName: "config-data") pod "253d6dae-a9bc-4da5-be6c-0e65311d34a2" (UID: "253d6dae-a9bc-4da5-be6c-0e65311d34a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.017589 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.017625 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfrc4\" (UniqueName: \"kubernetes.io/projected/253d6dae-a9bc-4da5-be6c-0e65311d34a2-kube-api-access-pfrc4\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.017643 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/253d6dae-a9bc-4da5-be6c-0e65311d34a2-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.017653 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.017667 4688 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/253d6dae-a9bc-4da5-be6c-0e65311d34a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.174172 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.185531 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-596f7f8496-72lx6"] Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.401207 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" path="/var/lib/kubelet/pods/253d6dae-a9bc-4da5-be6c-0e65311d34a2/volumes" Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.840089 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.840953 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-central-agent" containerID="cri-o://2a3d18b965114473d044a618e486a02e57613031dd297e6a780b4204a5af48ed" gracePeriod=30 Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.841499 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" containerID="cri-o://ce4b067d88280eee79125cdd3ff415b377d9202cab0b89ca2fd30e7cccf01ed0" gracePeriod=30 Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.841738 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-notification-agent" containerID="cri-o://37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7" gracePeriod=30 Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.841816 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="sg-core" containerID="cri-o://082c75db6aab9a4dae6032996e3712882be3e10e3b4573c57ce5d1421fee3265" gracePeriod=30 Oct 01 16:05:37 crc kubenswrapper[4688]: I1001 16:05:37.882059 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.006907 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6bbc5fdc47-8258p"] Oct 01 16:05:38 crc kubenswrapper[4688]: E1001 16:05:38.007223 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.007234 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api" Oct 01 16:05:38 crc kubenswrapper[4688]: E1001 16:05:38.011697 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api-log" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.011719 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api-log" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.011953 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.011969 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="253d6dae-a9bc-4da5-be6c-0e65311d34a2" containerName="barbican-api-log" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.012864 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.016928 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.017105 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.017227 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.040939 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tzm\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-kube-api-access-n8tzm\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.040987 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-public-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041027 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-config-data\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041050 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-run-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041079 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-combined-ca-bundle\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041092 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-internal-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041114 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-log-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.041174 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-etc-swift\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.079728 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bbc5fdc47-8258p"] Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144108 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-public-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144192 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-config-data\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144216 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-run-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144246 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-combined-ca-bundle\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144282 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-internal-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144306 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-log-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144370 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-etc-swift\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.144397 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tzm\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-kube-api-access-n8tzm\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.146370 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-log-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.149973 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/43929b9b-990a-4a85-9100-509d80bba3c1-run-httpd\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.150209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-config-data\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.151963 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-combined-ca-bundle\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.158487 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-internal-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.177156 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43929b9b-990a-4a85-9100-509d80bba3c1-public-tls-certs\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.178781 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-etc-swift\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.179436 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tzm\" (UniqueName: \"kubernetes.io/projected/43929b9b-990a-4a85-9100-509d80bba3c1-kube-api-access-n8tzm\") pod \"swift-proxy-6bbc5fdc47-8258p\" (UID: \"43929b9b-990a-4a85-9100-509d80bba3c1\") " pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.356583 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.926857 4688 generic.go:334] "Generic (PLEG): container finished" podID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerID="ce4b067d88280eee79125cdd3ff415b377d9202cab0b89ca2fd30e7cccf01ed0" exitCode=0 Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.927249 4688 generic.go:334] "Generic (PLEG): container finished" podID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerID="082c75db6aab9a4dae6032996e3712882be3e10e3b4573c57ce5d1421fee3265" exitCode=2 Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.927269 4688 generic.go:334] "Generic (PLEG): container finished" podID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerID="2a3d18b965114473d044a618e486a02e57613031dd297e6a780b4204a5af48ed" exitCode=0 Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.927302 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerDied","Data":"ce4b067d88280eee79125cdd3ff415b377d9202cab0b89ca2fd30e7cccf01ed0"} Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.927344 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerDied","Data":"082c75db6aab9a4dae6032996e3712882be3e10e3b4573c57ce5d1421fee3265"} Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.927367 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerDied","Data":"2a3d18b965114473d044a618e486a02e57613031dd297e6a780b4204a5af48ed"} Oct 01 16:05:38 crc kubenswrapper[4688]: I1001 16:05:38.936321 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bbc5fdc47-8258p"] Oct 01 16:05:38 crc kubenswrapper[4688]: W1001 16:05:38.948076 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43929b9b_990a_4a85_9100_509d80bba3c1.slice/crio-9b859c2ddea5d509a372e30425b3611f2bbc7cfc6413d1fa37db37474029cb42 WatchSource:0}: Error finding container 9b859c2ddea5d509a372e30425b3611f2bbc7cfc6413d1fa37db37474029cb42: Status 404 returned error can't find the container with id 9b859c2ddea5d509a372e30425b3611f2bbc7cfc6413d1fa37db37474029cb42 Oct 01 16:05:39 crc kubenswrapper[4688]: E1001 16:05:39.625666 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13321c8d_8fe9_4036_9061_0f5466c84b6c.slice/crio-37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13321c8d_8fe9_4036_9061_0f5466c84b6c.slice/crio-conmon-37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.938623 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bbc5fdc47-8258p" event={"ID":"43929b9b-990a-4a85-9100-509d80bba3c1","Type":"ContainerStarted","Data":"38fc0ea7aad7999e19cf4136d7837dccdf3f10cd33624306079dcc9ead41bba9"} Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.938673 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bbc5fdc47-8258p" event={"ID":"43929b9b-990a-4a85-9100-509d80bba3c1","Type":"ContainerStarted","Data":"7da668877f14accedcb806e097fde41a7f7dc9da57a3a32dd09aff757fe38653"} Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.938686 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bbc5fdc47-8258p" event={"ID":"43929b9b-990a-4a85-9100-509d80bba3c1","Type":"ContainerStarted","Data":"9b859c2ddea5d509a372e30425b3611f2bbc7cfc6413d1fa37db37474029cb42"} Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.940003 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.940060 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.944605 4688 generic.go:334] "Generic (PLEG): container finished" podID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerID="37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7" exitCode=0 Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.944654 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerDied","Data":"37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7"} Oct 01 16:05:39 crc kubenswrapper[4688]: I1001 16:05:39.969016 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6bbc5fdc47-8258p" podStartSLOduration=2.9689789429999998 podStartE2EDuration="2.968978943s" podCreationTimestamp="2025-10-01 16:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:39.957405801 +0000 UTC m=+1129.308045763" watchObservedRunningTime="2025-10-01 16:05:39.968978943 +0000 UTC m=+1129.319618905" Oct 01 16:05:41 crc kubenswrapper[4688]: I1001 16:05:41.323179 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:41 crc kubenswrapper[4688]: I1001 16:05:41.323999 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-httpd" containerID="cri-o://46257c1c7425d0ab02e789a29a47d108581911b07f7b67fed4e7e9bd27c4d1ef" gracePeriod=30 Oct 01 16:05:41 crc kubenswrapper[4688]: I1001 16:05:41.323698 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-log" containerID="cri-o://e7a26a406d05cbe87f75cf5049289106ab60f10b02d02313f13255906a60ef2f" gracePeriod=30 Oct 01 16:05:41 crc kubenswrapper[4688]: I1001 16:05:41.979382 4688 generic.go:334] "Generic (PLEG): container finished" podID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerID="e7a26a406d05cbe87f75cf5049289106ab60f10b02d02313f13255906a60ef2f" exitCode=143 Oct 01 16:05:41 crc kubenswrapper[4688]: I1001 16:05:41.980392 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerDied","Data":"e7a26a406d05cbe87f75cf5049289106ab60f10b02d02313f13255906a60ef2f"} Oct 01 16:05:42 crc kubenswrapper[4688]: I1001 16:05:42.890769 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-dn99k"] Oct 01 16:05:42 crc kubenswrapper[4688]: I1001 16:05:42.892033 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:42 crc kubenswrapper[4688]: I1001 16:05:42.910494 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dn99k"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.042121 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4766\" (UniqueName: \"kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766\") pod \"nova-api-db-create-dn99k\" (UID: \"f89bebef-c272-4141-9659-0621a89e1baa\") " pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.081631 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-hdr4l"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.082686 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.094571 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hdr4l"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.144368 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4766\" (UniqueName: \"kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766\") pod \"nova-api-db-create-dn99k\" (UID: \"f89bebef-c272-4141-9659-0621a89e1baa\") " pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.197676 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4766\" (UniqueName: \"kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766\") pod \"nova-api-db-create-dn99k\" (UID: \"f89bebef-c272-4141-9659-0621a89e1baa\") " pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.212546 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.218217 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-df88w"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.219713 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.245608 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrk8s\" (UniqueName: \"kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s\") pod \"nova-cell0-db-create-hdr4l\" (UID: \"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf\") " pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.247674 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-df88w"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.300348 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.300683 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-log" containerID="cri-o://669ce4b84dfa8d9deb9d930ec89a60575ab320d2060fc2071b9e14a15b19aa6e" gracePeriod=30 Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.301175 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-httpd" containerID="cri-o://60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f" gracePeriod=30 Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.311978 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.156:9292/healthcheck\": EOF" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.350454 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt8px\" (UniqueName: \"kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px\") pod \"nova-cell1-db-create-df88w\" (UID: \"c8dd8f9e-3618-40a4-a5a8-379142afdf65\") " pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.350684 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrk8s\" (UniqueName: \"kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s\") pod \"nova-cell0-db-create-hdr4l\" (UID: \"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf\") " pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.396381 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrk8s\" (UniqueName: \"kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s\") pod \"nova-cell0-db-create-hdr4l\" (UID: \"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf\") " pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.412187 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.452227 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt8px\" (UniqueName: \"kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px\") pod \"nova-cell1-db-create-df88w\" (UID: \"c8dd8f9e-3618-40a4-a5a8-379142afdf65\") " pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.474016 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt8px\" (UniqueName: \"kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px\") pod \"nova-cell1-db-create-df88w\" (UID: \"c8dd8f9e-3618-40a4-a5a8-379142afdf65\") " pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:43 crc kubenswrapper[4688]: I1001 16:05:43.557109 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:44 crc kubenswrapper[4688]: I1001 16:05:44.009678 4688 generic.go:334] "Generic (PLEG): container finished" podID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerID="669ce4b84dfa8d9deb9d930ec89a60575ab320d2060fc2071b9e14a15b19aa6e" exitCode=143 Oct 01 16:05:44 crc kubenswrapper[4688]: I1001 16:05:44.009732 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerDied","Data":"669ce4b84dfa8d9deb9d930ec89a60575ab320d2060fc2071b9e14a15b19aa6e"} Oct 01 16:05:45 crc kubenswrapper[4688]: I1001 16:05:45.023720 4688 generic.go:334] "Generic (PLEG): container finished" podID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerID="46257c1c7425d0ab02e789a29a47d108581911b07f7b67fed4e7e9bd27c4d1ef" exitCode=0 Oct 01 16:05:45 crc kubenswrapper[4688]: I1001 16:05:45.024760 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerDied","Data":"46257c1c7425d0ab02e789a29a47d108581911b07f7b67fed4e7e9bd27c4d1ef"} Oct 01 16:05:47 crc kubenswrapper[4688]: I1001 16:05:47.867477 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060004 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9267\" (UniqueName: \"kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060117 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060178 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060207 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060223 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060251 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.060278 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data\") pod \"13321c8d-8fe9-4036-9061-0f5466c84b6c\" (UID: \"13321c8d-8fe9-4036-9061-0f5466c84b6c\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.066537 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.069652 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.080968 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.081421 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13321c8d-8fe9-4036-9061-0f5466c84b6c","Type":"ContainerDied","Data":"a3c8d852dee5a9d4d93242e3a849ecc6ede8ab743e0b0c443fa5e338343b53d2"} Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.081459 4688 scope.go:117] "RemoveContainer" containerID="ce4b067d88280eee79125cdd3ff415b377d9202cab0b89ca2fd30e7cccf01ed0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.081588 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.100516 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts" (OuterVolumeSpecName: "scripts") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.142375 4688 scope.go:117] "RemoveContainer" containerID="082c75db6aab9a4dae6032996e3712882be3e10e3b4573c57ce5d1421fee3265" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.158485 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267" (OuterVolumeSpecName: "kube-api-access-z9267") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "kube-api-access-z9267". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.171457 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.171546 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.171559 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13321c8d-8fe9-4036-9061-0f5466c84b6c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.171576 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9267\" (UniqueName: \"kubernetes.io/projected/13321c8d-8fe9-4036-9061-0f5466c84b6c-kube-api-access-z9267\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.184242 4688 scope.go:117] "RemoveContainer" containerID="37b3a7c75fd729906fba7ded5debb535dbe644f558d7be594447a7e4def0daf7" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.188064 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.617459293 podStartE2EDuration="16.18804626s" podCreationTimestamp="2025-10-01 16:05:32 +0000 UTC" firstStartedPulling="2025-10-01 16:05:33.122692781 +0000 UTC m=+1122.473332733" lastFinishedPulling="2025-10-01 16:05:47.693279738 +0000 UTC m=+1137.043919700" observedRunningTime="2025-10-01 16:05:48.184572613 +0000 UTC m=+1137.535212575" watchObservedRunningTime="2025-10-01 16:05:48.18804626 +0000 UTC m=+1137.538686222" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.217153 4688 scope.go:117] "RemoveContainer" containerID="2a3d18b965114473d044a618e486a02e57613031dd297e6a780b4204a5af48ed" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.239287 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274494 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274667 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274709 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274736 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274771 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274801 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274854 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsw4d\" (UniqueName: \"kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.274889 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle\") pod \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\" (UID: \"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e\") " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.275228 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.283614 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.287496 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs" (OuterVolumeSpecName: "logs") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.293314 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts" (OuterVolumeSpecName: "scripts") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.293508 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.296459 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data" (OuterVolumeSpecName: "config-data") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.302746 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d" (OuterVolumeSpecName: "kube-api-access-nsw4d") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "kube-api-access-nsw4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.383461 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13321c8d-8fe9-4036-9061-0f5466c84b6c" (UID: "13321c8d-8fe9-4036-9061-0f5466c84b6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.383941 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsw4d\" (UniqueName: \"kubernetes.io/projected/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-kube-api-access-nsw4d\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.383973 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.383985 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.384184 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13321c8d-8fe9-4036-9061-0f5466c84b6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.384195 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.384220 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.384229 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.394465 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.425264 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.434215 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.474904 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-df88w"] Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.492109 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.508150 4688 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.509337 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.511479 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data" (OuterVolumeSpecName: "config-data") pod "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" (UID: "f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.515334 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.555935 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bbc5fdc47-8258p" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.590493 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dn99k"] Oct 01 16:05:48 crc kubenswrapper[4688]: W1001 16:05:48.594248 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf89bebef_c272_4141_9659_0621a89e1baa.slice/crio-70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a WatchSource:0}: Error finding container 70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a: Status 404 returned error can't find the container with id 70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.611187 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.620586 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hdr4l"] Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.759143 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.777473 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.811581 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.811991 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-central-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812008 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-central-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.812024 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-log" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812031 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-log" Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.812055 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812061 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.812073 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="sg-core" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812078 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="sg-core" Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.812086 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812092 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: E1001 16:05:48.812104 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-notification-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812110 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-notification-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812289 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="sg-core" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812305 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-log" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812319 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-notification-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812332 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812344 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="ceilometer-central-agent" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.812352 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" containerName="glance-httpd" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.813919 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.822930 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.823139 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.831005 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931140 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931172 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931189 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931430 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931458 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-485fl\" (UniqueName: \"kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931651 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:48 crc kubenswrapper[4688]: I1001 16:05:48.931705 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.033204 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.034885 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.034909 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.034988 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.035018 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-485fl\" (UniqueName: \"kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.035086 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.035102 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.035427 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.036180 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.042094 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.042187 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.043215 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.046013 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.053588 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-485fl\" (UniqueName: \"kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl\") pod \"ceilometer-0\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.121443 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a309f66f-cc9d-40e7-9acc-00881395f4c9","Type":"ContainerStarted","Data":"65f5a3f397f1aa6a71a0ca3164aa5ebe7a691da92c778e9ce41ffa8ec7702f2e"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.123854 4688 generic.go:334] "Generic (PLEG): container finished" podID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerID="674bcb3920fc3555ff6521c9241431dd30559cff036703357069642077847a95" exitCode=137 Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.123900 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-555656bf5b-xlntq" event={"ID":"37923e92-dbcc-41a1-8d2f-89d8de59959e","Type":"ContainerDied","Data":"674bcb3920fc3555ff6521c9241431dd30559cff036703357069642077847a95"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.126559 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdr4l" event={"ID":"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf","Type":"ContainerStarted","Data":"d733f7af370aed949f3e5ed98c0c6853d1d7bf6928bf535783077588aaafb243"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.126595 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdr4l" event={"ID":"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf","Type":"ContainerStarted","Data":"ba4591140d0a69d295b57d19ed495d7746e7eb0e44fc4133f822a19b4efa3bd6"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.129137 4688 generic.go:334] "Generic (PLEG): container finished" podID="c8dd8f9e-3618-40a4-a5a8-379142afdf65" containerID="be433fced6fadcf1fae67dad74c865a4e88c858e8b4d13d20623521c4a9fdf47" exitCode=0 Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.129188 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df88w" event={"ID":"c8dd8f9e-3618-40a4-a5a8-379142afdf65","Type":"ContainerDied","Data":"be433fced6fadcf1fae67dad74c865a4e88c858e8b4d13d20623521c4a9fdf47"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.129208 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df88w" event={"ID":"c8dd8f9e-3618-40a4-a5a8-379142afdf65","Type":"ContainerStarted","Data":"be2c01b4d3fee590d2ab6810eca7a44e2a9cec7d8319d12b2312619942c6097e"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.132314 4688 generic.go:334] "Generic (PLEG): container finished" podID="85db256a-98d5-4bca-82ac-098d49b243e8" containerID="db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2" exitCode=137 Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.132362 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerDied","Data":"db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.136597 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e","Type":"ContainerDied","Data":"8c0c2ff4a3dd5be834a2e028a22c76a76b5ef50aa7d3a8867165c97af80d3210"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.136634 4688 scope.go:117] "RemoveContainer" containerID="46257c1c7425d0ab02e789a29a47d108581911b07f7b67fed4e7e9bd27c4d1ef" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.136742 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.147072 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dn99k" event={"ID":"f89bebef-c272-4141-9659-0621a89e1baa","Type":"ContainerStarted","Data":"26c7cb4e6b855793e97c207822f4a551fbb42b3d8e733269b88b27fcb86356b0"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.147105 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dn99k" event={"ID":"f89bebef-c272-4141-9659-0621a89e1baa","Type":"ContainerStarted","Data":"70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a"} Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.154368 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-hdr4l" podStartSLOduration=6.154345088 podStartE2EDuration="6.154345088s" podCreationTimestamp="2025-10-01 16:05:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:49.147173459 +0000 UTC m=+1138.497813421" watchObservedRunningTime="2025-10-01 16:05:49.154345088 +0000 UTC m=+1138.504985050" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.167397 4688 scope.go:117] "RemoveContainer" containerID="e7a26a406d05cbe87f75cf5049289106ab60f10b02d02313f13255906a60ef2f" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.185372 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-dn99k" podStartSLOduration=7.1853513509999996 podStartE2EDuration="7.185351351s" podCreationTimestamp="2025-10-01 16:05:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:49.17345865 +0000 UTC m=+1138.524098612" watchObservedRunningTime="2025-10-01 16:05:49.185351351 +0000 UTC m=+1138.535991313" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.221578 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.233244 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.240278 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.251294 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.252858 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.255628 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.259362 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.265479 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.340666 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.340768 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.340974 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.341020 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.341174 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.341200 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-logs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.341248 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9l2t\" (UniqueName: \"kubernetes.io/projected/32ec1a41-4380-4f29-a56d-dd2d722999d6-kube-api-access-c9l2t\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.341294 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.396839 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" path="/var/lib/kubelet/pods/13321c8d-8fe9-4036-9061-0f5466c84b6c/volumes" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.397663 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e" path="/var/lib/kubelet/pods/f24d7e9c-1e3f-4a7b-b88e-234b98f4ac2e/volumes" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442546 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442616 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442640 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442927 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442946 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-logs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442973 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9l2t\" (UniqueName: \"kubernetes.io/projected/32ec1a41-4380-4f29-a56d-dd2d722999d6-kube-api-access-c9l2t\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.442998 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.443030 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.444456 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.458054 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.463341 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32ec1a41-4380-4f29-a56d-dd2d722999d6-logs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.467502 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.470079 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.470366 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.470906 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ec1a41-4380-4f29-a56d-dd2d722999d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.478250 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9l2t\" (UniqueName: \"kubernetes.io/projected/32ec1a41-4380-4f29-a56d-dd2d722999d6-kube-api-access-c9l2t\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.483984 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"32ec1a41-4380-4f29-a56d-dd2d722999d6\") " pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.618017 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 16:05:49 crc kubenswrapper[4688]: I1001 16:05:49.775236 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:50 crc kubenswrapper[4688]: E1001 16:05:50.056032 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod116546ec_6ffa_4b68_8dd1_c455793fe29f.slice/crio-60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f.scope\": RecentStats: unable to find data in memory cache]" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.159297 4688 generic.go:334] "Generic (PLEG): container finished" podID="f89bebef-c272-4141-9659-0621a89e1baa" containerID="26c7cb4e6b855793e97c207822f4a551fbb42b3d8e733269b88b27fcb86356b0" exitCode=0 Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.159352 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dn99k" event={"ID":"f89bebef-c272-4141-9659-0621a89e1baa","Type":"ContainerDied","Data":"26c7cb4e6b855793e97c207822f4a551fbb42b3d8e733269b88b27fcb86356b0"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.161518 4688 generic.go:334] "Generic (PLEG): container finished" podID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerID="60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f" exitCode=0 Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.161586 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerDied","Data":"60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.163725 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-555656bf5b-xlntq" event={"ID":"37923e92-dbcc-41a1-8d2f-89d8de59959e","Type":"ContainerStarted","Data":"deab0124a0e443133dbbb1a6975c4dd4a3c0d59d9d4007729706031e440cbb65"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.165812 4688 generic.go:334] "Generic (PLEG): container finished" podID="af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" containerID="d733f7af370aed949f3e5ed98c0c6853d1d7bf6928bf535783077588aaafb243" exitCode=0 Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.165855 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdr4l" event={"ID":"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf","Type":"ContainerDied","Data":"d733f7af370aed949f3e5ed98c0c6853d1d7bf6928bf535783077588aaafb243"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.166834 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerStarted","Data":"cf692657c1341fb3177a7294f7575aa4c801e512bfe5e9a8b4981ac0c4dfa95c"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.172476 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerStarted","Data":"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4"} Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.314138 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.384134 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.384337 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.384499 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.384593 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.384689 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.385521 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww5wz\" (UniqueName: \"kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.385611 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.385714 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle\") pod \"116546ec-6ffa-4b68-8dd1-c455793fe29f\" (UID: \"116546ec-6ffa-4b68-8dd1-c455793fe29f\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.388756 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.389915 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs" (OuterVolumeSpecName: "logs") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.391769 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts" (OuterVolumeSpecName: "scripts") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.393236 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.394728 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz" (OuterVolumeSpecName: "kube-api-access-ww5wz") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "kube-api-access-ww5wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.424603 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492175 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492568 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492644 4688 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492715 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/116546ec-6ffa-4b68-8dd1-c455793fe29f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492768 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww5wz\" (UniqueName: \"kubernetes.io/projected/116546ec-6ffa-4b68-8dd1-c455793fe29f-kube-api-access-ww5wz\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492820 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.492877 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.514927 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data" (OuterVolumeSpecName: "config-data") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.515093 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "116546ec-6ffa-4b68-8dd1-c455793fe29f" (UID: "116546ec-6ffa-4b68-8dd1-c455793fe29f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.576093 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.605757 4688 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.605785 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.605794 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/116546ec-6ffa-4b68-8dd1-c455793fe29f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.689845 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.808320 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt8px\" (UniqueName: \"kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px\") pod \"c8dd8f9e-3618-40a4-a5a8-379142afdf65\" (UID: \"c8dd8f9e-3618-40a4-a5a8-379142afdf65\") " Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.823112 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px" (OuterVolumeSpecName: "kube-api-access-rt8px") pod "c8dd8f9e-3618-40a4-a5a8-379142afdf65" (UID: "c8dd8f9e-3618-40a4-a5a8-379142afdf65"). InnerVolumeSpecName "kube-api-access-rt8px". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:50 crc kubenswrapper[4688]: I1001 16:05:50.912145 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt8px\" (UniqueName: \"kubernetes.io/projected/c8dd8f9e-3618-40a4-a5a8-379142afdf65-kube-api-access-rt8px\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.192283 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-df88w" event={"ID":"c8dd8f9e-3618-40a4-a5a8-379142afdf65","Type":"ContainerDied","Data":"be2c01b4d3fee590d2ab6810eca7a44e2a9cec7d8319d12b2312619942c6097e"} Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.192347 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be2c01b4d3fee590d2ab6810eca7a44e2a9cec7d8319d12b2312619942c6097e" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.192310 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-df88w" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.194251 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"32ec1a41-4380-4f29-a56d-dd2d722999d6","Type":"ContainerStarted","Data":"3a642e3825a2ddfb4ae5c30bd1fa71b845437eeac62d1926c2ef2b1b674963a2"} Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.195737 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"116546ec-6ffa-4b68-8dd1-c455793fe29f","Type":"ContainerDied","Data":"13d7ca8afb588a2930fb85629b299a193fbcd9d5cd0e456cc1c1a02defa60648"} Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.195790 4688 scope.go:117] "RemoveContainer" containerID="60960420f6eaf0f33629e2c911c73337b6089965fc8d99a3879dfb749e5a5b3f" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.195808 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.235444 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.249132 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.266587 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:51 crc kubenswrapper[4688]: E1001 16:05:51.267204 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-httpd" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.267296 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-httpd" Oct 01 16:05:51 crc kubenswrapper[4688]: E1001 16:05:51.267378 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dd8f9e-3618-40a4-a5a8-379142afdf65" containerName="mariadb-database-create" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.267441 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dd8f9e-3618-40a4-a5a8-379142afdf65" containerName="mariadb-database-create" Oct 01 16:05:51 crc kubenswrapper[4688]: E1001 16:05:51.267532 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-log" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.267612 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-log" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.267856 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-httpd" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.267947 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8dd8f9e-3618-40a4-a5a8-379142afdf65" containerName="mariadb-database-create" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.268034 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" containerName="glance-log" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.274179 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.284940 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.286060 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.292690 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318280 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318401 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318438 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318479 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318517 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318625 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g46jn\" (UniqueName: \"kubernetes.io/projected/ab7d6e5f-ceab-400c-858e-2059dc492923-kube-api-access-g46jn\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318669 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.318744 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.391763 4688 scope.go:117] "RemoveContainer" containerID="669ce4b84dfa8d9deb9d930ec89a60575ab320d2060fc2071b9e14a15b19aa6e" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.426668 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.429583 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.429772 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g46jn\" (UniqueName: \"kubernetes.io/projected/ab7d6e5f-ceab-400c-858e-2059dc492923-kube-api-access-g46jn\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.429886 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.434630 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.437102 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-logs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.438920 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.439145 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.439352 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.439461 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.441547 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ab7d6e5f-ceab-400c-858e-2059dc492923-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.446485 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.447178 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.447388 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.457005 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.457402 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.462569 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116546ec-6ffa-4b68-8dd1-c455793fe29f" path="/var/lib/kubelet/pods/116546ec-6ffa-4b68-8dd1-c455793fe29f/volumes" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.467608 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab7d6e5f-ceab-400c-858e-2059dc492923-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.580566 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g46jn\" (UniqueName: \"kubernetes.io/projected/ab7d6e5f-ceab-400c-858e-2059dc492923-kube-api-access-g46jn\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.758935 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"ab7d6e5f-ceab-400c-858e-2059dc492923\") " pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.906937 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 16:05:51 crc kubenswrapper[4688]: I1001 16:05:51.910042 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.040615 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrk8s\" (UniqueName: \"kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s\") pod \"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf\" (UID: \"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf\") " Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.053875 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s" (OuterVolumeSpecName: "kube-api-access-rrk8s") pod "af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" (UID: "af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf"). InnerVolumeSpecName "kube-api-access-rrk8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.143827 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrk8s\" (UniqueName: \"kubernetes.io/projected/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf-kube-api-access-rrk8s\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.238250 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdr4l" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.238281 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdr4l" event={"ID":"af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf","Type":"ContainerDied","Data":"ba4591140d0a69d295b57d19ed495d7746e7eb0e44fc4133f822a19b4efa3bd6"} Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.240469 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba4591140d0a69d295b57d19ed495d7746e7eb0e44fc4133f822a19b4efa3bd6" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.265772 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerStarted","Data":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.265939 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.349796 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4766\" (UniqueName: \"kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766\") pod \"f89bebef-c272-4141-9659-0621a89e1baa\" (UID: \"f89bebef-c272-4141-9659-0621a89e1baa\") " Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.360709 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766" (OuterVolumeSpecName: "kube-api-access-f4766") pod "f89bebef-c272-4141-9659-0621a89e1baa" (UID: "f89bebef-c272-4141-9659-0621a89e1baa"). InnerVolumeSpecName "kube-api-access-f4766". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.451934 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4766\" (UniqueName: \"kubernetes.io/projected/f89bebef-c272-4141-9659-0621a89e1baa-kube-api-access-f4766\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:52 crc kubenswrapper[4688]: I1001 16:05:52.816592 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 16:05:52 crc kubenswrapper[4688]: W1001 16:05:52.846163 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab7d6e5f_ceab_400c_858e_2059dc492923.slice/crio-a4942decc4f79407dbcd1616ad200b74667ebaabe1f321d180895588b557f5c0 WatchSource:0}: Error finding container a4942decc4f79407dbcd1616ad200b74667ebaabe1f321d180895588b557f5c0: Status 404 returned error can't find the container with id a4942decc4f79407dbcd1616ad200b74667ebaabe1f321d180895588b557f5c0 Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.281329 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dn99k" event={"ID":"f89bebef-c272-4141-9659-0621a89e1baa","Type":"ContainerDied","Data":"70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a"} Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.281686 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70c2d28c1317b80b9aec1f0cd331489eb54e4a73d29f2a08383f8dcad03bf27a" Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.281747 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dn99k" Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.303880 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d6e5f-ceab-400c-858e-2059dc492923","Type":"ContainerStarted","Data":"a4942decc4f79407dbcd1616ad200b74667ebaabe1f321d180895588b557f5c0"} Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.314873 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerStarted","Data":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.320063 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"32ec1a41-4380-4f29-a56d-dd2d722999d6","Type":"ContainerStarted","Data":"fcf7270ba60dc9475f04a188ae63f98986dfc3a52a3ba1bfc76c06c2544829f3"} Oct 01 16:05:53 crc kubenswrapper[4688]: I1001 16:05:53.438310 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:05:54 crc kubenswrapper[4688]: I1001 16:05:54.337732 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d6e5f-ceab-400c-858e-2059dc492923","Type":"ContainerStarted","Data":"de35b6aee83c4e6930ef0ea4098d53ab8a52deb164629adfa41e65955d1efb06"} Oct 01 16:05:54 crc kubenswrapper[4688]: I1001 16:05:54.344201 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerStarted","Data":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} Oct 01 16:05:54 crc kubenswrapper[4688]: I1001 16:05:54.348653 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"32ec1a41-4380-4f29-a56d-dd2d722999d6","Type":"ContainerStarted","Data":"15c08fce795af5d7c38b8802da26a96a32f71074ac26c9d23f3913bce35da70c"} Oct 01 16:05:54 crc kubenswrapper[4688]: I1001 16:05:54.375265 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.37523286 podStartE2EDuration="5.37523286s" podCreationTimestamp="2025-10-01 16:05:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:54.370409676 +0000 UTC m=+1143.721049648" watchObservedRunningTime="2025-10-01 16:05:54.37523286 +0000 UTC m=+1143.725872902" Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.052601 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.053141 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.053185 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.053852 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.053906 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18" gracePeriod=600 Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.362013 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18" exitCode=0 Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.362275 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18"} Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.362381 4688 scope.go:117] "RemoveContainer" containerID="1bdf3b65d7c85ad934f4fb82f8d9db9baafd9b6c4e5793f1558b36f7ac094ddd" Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.366600 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ab7d6e5f-ceab-400c-858e-2059dc492923","Type":"ContainerStarted","Data":"30ecabf0a9c0d2700defbfd2e5316fdea266b98f10649723dfa46aaf3c9a0dc3"} Oct 01 16:05:55 crc kubenswrapper[4688]: I1001 16:05:55.392938 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.392922066 podStartE2EDuration="4.392922066s" podCreationTimestamp="2025-10-01 16:05:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:05:55.392817734 +0000 UTC m=+1144.743457696" watchObservedRunningTime="2025-10-01 16:05:55.392922066 +0000 UTC m=+1144.743562028" Oct 01 16:05:56 crc kubenswrapper[4688]: I1001 16:05:56.376991 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c"} Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.398267 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerStarted","Data":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.398735 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-central-agent" containerID="cri-o://99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" gracePeriod=30 Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.398993 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.399020 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="proxy-httpd" containerID="cri-o://d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" gracePeriod=30 Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.399144 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="sg-core" containerID="cri-o://70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" gracePeriod=30 Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.399188 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-notification-agent" containerID="cri-o://79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" gracePeriod=30 Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.427016 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.828249209 podStartE2EDuration="10.426994001s" podCreationTimestamp="2025-10-01 16:05:48 +0000 UTC" firstStartedPulling="2025-10-01 16:05:49.791387758 +0000 UTC m=+1139.142027720" lastFinishedPulling="2025-10-01 16:05:57.39013256 +0000 UTC m=+1146.740772512" observedRunningTime="2025-10-01 16:05:58.42120632 +0000 UTC m=+1147.771846282" watchObservedRunningTime="2025-10-01 16:05:58.426994001 +0000 UTC m=+1147.777633963" Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.538235 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.538294 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.588133 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:05:58 crc kubenswrapper[4688]: I1001 16:05:58.588373 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.406236 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407117 4688 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" exitCode=0 Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407137 4688 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" exitCode=2 Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407145 4688 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" exitCode=0 Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407151 4688 generic.go:334] "Generic (PLEG): container finished" podID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" exitCode=0 Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407855 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerDied","Data":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407885 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerDied","Data":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407896 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerDied","Data":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407904 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerDied","Data":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407914 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6eef09a3-3e52-4735-92ef-5ddb81c6d86b","Type":"ContainerDied","Data":"cf692657c1341fb3177a7294f7575aa4c801e512bfe5e9a8b4981ac0c4dfa95c"} Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.407929 4688 scope.go:117] "RemoveContainer" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.447696 4688 scope.go:117] "RemoveContainer" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.495589 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.495961 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.496019 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.496052 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.496090 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.496190 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-485fl\" (UniqueName: \"kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.496274 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle\") pod \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\" (UID: \"6eef09a3-3e52-4735-92ef-5ddb81c6d86b\") " Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.497117 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.500675 4688 scope.go:117] "RemoveContainer" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.501492 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.514510 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl" (OuterVolumeSpecName: "kube-api-access-485fl") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "kube-api-access-485fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.517509 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts" (OuterVolumeSpecName: "scripts") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.561748 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.603556 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-485fl\" (UniqueName: \"kubernetes.io/projected/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-kube-api-access-485fl\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.603588 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.603597 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.603605 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.603613 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.619396 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.619447 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.656045 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data" (OuterVolumeSpecName: "config-data") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.666487 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6eef09a3-3e52-4735-92ef-5ddb81c6d86b" (UID: "6eef09a3-3e52-4735-92ef-5ddb81c6d86b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.698641 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.705270 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.705300 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eef09a3-3e52-4735-92ef-5ddb81c6d86b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.706383 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.772491 4688 scope.go:117] "RemoveContainer" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.790020 4688 scope.go:117] "RemoveContainer" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: E1001 16:05:59.790404 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": container with ID starting with d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44 not found: ID does not exist" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.790435 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} err="failed to get container status \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": rpc error: code = NotFound desc = could not find container \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": container with ID starting with d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.790454 4688 scope.go:117] "RemoveContainer" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: E1001 16:05:59.790913 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": container with ID starting with 70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d not found: ID does not exist" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.790972 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} err="failed to get container status \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": rpc error: code = NotFound desc = could not find container \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": container with ID starting with 70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.791005 4688 scope.go:117] "RemoveContainer" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: E1001 16:05:59.791395 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": container with ID starting with 79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c not found: ID does not exist" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.791421 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} err="failed to get container status \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": rpc error: code = NotFound desc = could not find container \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": container with ID starting with 79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.791437 4688 scope.go:117] "RemoveContainer" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: E1001 16:05:59.791908 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": container with ID starting with 99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132 not found: ID does not exist" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.791950 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} err="failed to get container status \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": rpc error: code = NotFound desc = could not find container \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": container with ID starting with 99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.791978 4688 scope.go:117] "RemoveContainer" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.792349 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} err="failed to get container status \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": rpc error: code = NotFound desc = could not find container \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": container with ID starting with d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.792406 4688 scope.go:117] "RemoveContainer" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.792739 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} err="failed to get container status \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": rpc error: code = NotFound desc = could not find container \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": container with ID starting with 70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.792763 4688 scope.go:117] "RemoveContainer" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793029 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} err="failed to get container status \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": rpc error: code = NotFound desc = could not find container \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": container with ID starting with 79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793089 4688 scope.go:117] "RemoveContainer" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793360 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} err="failed to get container status \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": rpc error: code = NotFound desc = could not find container \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": container with ID starting with 99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793385 4688 scope.go:117] "RemoveContainer" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793617 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} err="failed to get container status \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": rpc error: code = NotFound desc = could not find container \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": container with ID starting with d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793648 4688 scope.go:117] "RemoveContainer" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793840 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} err="failed to get container status \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": rpc error: code = NotFound desc = could not find container \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": container with ID starting with 70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.793858 4688 scope.go:117] "RemoveContainer" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794127 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} err="failed to get container status \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": rpc error: code = NotFound desc = could not find container \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": container with ID starting with 79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794147 4688 scope.go:117] "RemoveContainer" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794321 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} err="failed to get container status \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": rpc error: code = NotFound desc = could not find container \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": container with ID starting with 99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794342 4688 scope.go:117] "RemoveContainer" containerID="d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794595 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44"} err="failed to get container status \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": rpc error: code = NotFound desc = could not find container \"d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44\": container with ID starting with d3ce99c3c6825150e46c136962055e02881868f27c568b3f0c080c07c8305f44 not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794613 4688 scope.go:117] "RemoveContainer" containerID="70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794865 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d"} err="failed to get container status \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": rpc error: code = NotFound desc = could not find container \"70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d\": container with ID starting with 70d95394f04318d445e1e37f92df08eaf984907ca43fac3e0bf948862f192d8d not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.794887 4688 scope.go:117] "RemoveContainer" containerID="79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.795108 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c"} err="failed to get container status \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": rpc error: code = NotFound desc = could not find container \"79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c\": container with ID starting with 79694621fdd1ea09b18eddf92da3d443dceefb933514d77bfe7e3af770bce12c not found: ID does not exist" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.795127 4688 scope.go:117] "RemoveContainer" containerID="99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132" Oct 01 16:05:59 crc kubenswrapper[4688]: I1001 16:05:59.795359 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132"} err="failed to get container status \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": rpc error: code = NotFound desc = could not find container \"99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132\": container with ID starting with 99c9043c7fb29022e4c5d879da17209ad90263211f590d734ce0922705726132 not found: ID does not exist" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.416993 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.417984 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.418081 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.451755 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.492680 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.508705 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.515941 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="proxy-httpd" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.516200 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="proxy-httpd" Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.516339 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.516416 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.516538 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-notification-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.516621 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-notification-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.516729 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="sg-core" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.516816 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="sg-core" Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.516927 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-central-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.517023 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-central-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: E1001 16:06:00.517147 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89bebef-c272-4141-9659-0621a89e1baa" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.517227 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89bebef-c272-4141-9659-0621a89e1baa" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.517943 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-central-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.518054 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="sg-core" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.518145 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="proxy-httpd" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.518244 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" containerName="ceilometer-notification-agent" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.518343 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89bebef-c272-4141-9659-0621a89e1baa" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.518440 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" containerName="mariadb-database-create" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.529165 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.531115 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.532591 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.532788 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.731778 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.732329 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.732471 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.732643 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nmsq\" (UniqueName: \"kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.732761 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.732973 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.733111 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.833930 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834019 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834058 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834090 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nmsq\" (UniqueName: \"kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834118 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834182 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.834227 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.835106 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.835160 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.839997 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.843154 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.848157 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.852186 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.870996 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nmsq\" (UniqueName: \"kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq\") pod \"ceilometer-0\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " pod="openstack/ceilometer-0" Oct 01 16:06:00 crc kubenswrapper[4688]: I1001 16:06:00.889880 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.396025 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eef09a3-3e52-4735-92ef-5ddb81c6d86b" path="/var/lib/kubelet/pods/6eef09a3-3e52-4735-92ef-5ddb81c6d86b/volumes" Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.396988 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.427597 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerStarted","Data":"5160616d9e12c7eac7a0bf15c21c7908ccc5b218dd07e929c848f0ea833fc69c"} Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.907453 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.907814 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.978958 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:01 crc kubenswrapper[4688]: I1001 16:06:01.979960 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.436905 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerStarted","Data":"e7c6acb51677d71c306f88abf20ac4f22d30b4bdd928e30314cba5d0988b0be8"} Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.437234 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.437253 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.436968 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.437273 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.930040 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:06:02 crc kubenswrapper[4688]: I1001 16:06:02.933083 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.134105 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7327-account-create-7xldq"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.135517 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.137315 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.143988 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7327-account-create-7xldq"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.285339 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w56gb\" (UniqueName: \"kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb\") pod \"nova-api-7327-account-create-7xldq\" (UID: \"cfe56a6d-2d84-449b-b5a8-116c5a9891b1\") " pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.341757 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-71b5-account-create-v7977"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.342963 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.351924 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.355084 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-71b5-account-create-v7977"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.394242 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w56gb\" (UniqueName: \"kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb\") pod \"nova-api-7327-account-create-7xldq\" (UID: \"cfe56a6d-2d84-449b-b5a8-116c5a9891b1\") " pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.436920 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w56gb\" (UniqueName: \"kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb\") pod \"nova-api-7327-account-create-7xldq\" (UID: \"cfe56a6d-2d84-449b-b5a8-116c5a9891b1\") " pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.459298 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.469595 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerStarted","Data":"54fc175da73a5d3a8cf1f937e6427e1c6a78aa42ecc69f1e754976144f8ec0ed"} Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.496465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mndrw\" (UniqueName: \"kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw\") pod \"nova-cell0-71b5-account-create-v7977\" (UID: \"13c68175-ca4c-4060-ac38-59f953323712\") " pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.551930 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9cc9-account-create-pxjr8"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.553247 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.562292 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9cc9-account-create-pxjr8"] Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.562556 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.601486 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mndrw\" (UniqueName: \"kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw\") pod \"nova-cell0-71b5-account-create-v7977\" (UID: \"13c68175-ca4c-4060-ac38-59f953323712\") " pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.648863 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mndrw\" (UniqueName: \"kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw\") pod \"nova-cell0-71b5-account-create-v7977\" (UID: \"13c68175-ca4c-4060-ac38-59f953323712\") " pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.693982 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.702665 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m97k\" (UniqueName: \"kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k\") pod \"nova-cell1-9cc9-account-create-pxjr8\" (UID: \"635e82b7-cdaf-4cc7-b1a9-6ed165196067\") " pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.805352 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m97k\" (UniqueName: \"kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k\") pod \"nova-cell1-9cc9-account-create-pxjr8\" (UID: \"635e82b7-cdaf-4cc7-b1a9-6ed165196067\") " pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.841804 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m97k\" (UniqueName: \"kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k\") pod \"nova-cell1-9cc9-account-create-pxjr8\" (UID: \"635e82b7-cdaf-4cc7-b1a9-6ed165196067\") " pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:03 crc kubenswrapper[4688]: I1001 16:06:03.900434 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.086374 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7327-account-create-7xldq"] Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.258661 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-71b5-account-create-v7977"] Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.492360 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerStarted","Data":"3e4014a2d4fafe2104bcb7d74cdbefbffb133cd3dbbdc397f09315b269e9b1f3"} Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.499086 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7327-account-create-7xldq" event={"ID":"cfe56a6d-2d84-449b-b5a8-116c5a9891b1","Type":"ContainerStarted","Data":"e4dac6c5577c30afe60971b6dd2e7ec7a374ffeec22889c8ea8806dba1a45c21"} Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.499305 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7327-account-create-7xldq" event={"ID":"cfe56a6d-2d84-449b-b5a8-116c5a9891b1","Type":"ContainerStarted","Data":"9d28969982b15f1930156de42aafec761577ee6e71cede772550721f0c3b7606"} Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.502155 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-71b5-account-create-v7977" event={"ID":"13c68175-ca4c-4060-ac38-59f953323712","Type":"ContainerStarted","Data":"d07f0c082806374cda394c9855359fc387ecc8ec85feb4241255c620c336239f"} Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.522648 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-7327-account-create-7xldq" podStartSLOduration=1.522631433 podStartE2EDuration="1.522631433s" podCreationTimestamp="2025-10-01 16:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:04.520639638 +0000 UTC m=+1153.871279600" watchObservedRunningTime="2025-10-01 16:06:04.522631433 +0000 UTC m=+1153.873271395" Oct 01 16:06:04 crc kubenswrapper[4688]: I1001 16:06:04.598199 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9cc9-account-create-pxjr8"] Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.334032 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.334390 4688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.349496 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.523341 4688 generic.go:334] "Generic (PLEG): container finished" podID="635e82b7-cdaf-4cc7-b1a9-6ed165196067" containerID="98b4a26ee7f4613dbd0e313f650a37d1cdb03864dd603b409f71aa9c8b587ffa" exitCode=0 Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.523437 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" event={"ID":"635e82b7-cdaf-4cc7-b1a9-6ed165196067","Type":"ContainerDied","Data":"98b4a26ee7f4613dbd0e313f650a37d1cdb03864dd603b409f71aa9c8b587ffa"} Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.523470 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" event={"ID":"635e82b7-cdaf-4cc7-b1a9-6ed165196067","Type":"ContainerStarted","Data":"ea31d90274fac3ae3e4e8b2bbdf85c3f722e0807e8762f9bbfe7f5ed2bbc091a"} Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.529149 4688 generic.go:334] "Generic (PLEG): container finished" podID="13c68175-ca4c-4060-ac38-59f953323712" containerID="e6b15d46587538238a5d3597e85aa02d88c3f5042165db6bc8c89142e9e8120b" exitCode=0 Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.529430 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-71b5-account-create-v7977" event={"ID":"13c68175-ca4c-4060-ac38-59f953323712","Type":"ContainerDied","Data":"e6b15d46587538238a5d3597e85aa02d88c3f5042165db6bc8c89142e9e8120b"} Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.531939 4688 generic.go:334] "Generic (PLEG): container finished" podID="cfe56a6d-2d84-449b-b5a8-116c5a9891b1" containerID="e4dac6c5577c30afe60971b6dd2e7ec7a374ffeec22889c8ea8806dba1a45c21" exitCode=0 Oct 01 16:06:05 crc kubenswrapper[4688]: I1001 16:06:05.533513 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7327-account-create-7xldq" event={"ID":"cfe56a6d-2d84-449b-b5a8-116c5a9891b1","Type":"ContainerDied","Data":"e4dac6c5577c30afe60971b6dd2e7ec7a374ffeec22889c8ea8806dba1a45c21"} Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.111632 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.220274 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.225860 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.235983 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w56gb\" (UniqueName: \"kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb\") pod \"cfe56a6d-2d84-449b-b5a8-116c5a9891b1\" (UID: \"cfe56a6d-2d84-449b-b5a8-116c5a9891b1\") " Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.242723 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb" (OuterVolumeSpecName: "kube-api-access-w56gb") pod "cfe56a6d-2d84-449b-b5a8-116c5a9891b1" (UID: "cfe56a6d-2d84-449b-b5a8-116c5a9891b1"). InnerVolumeSpecName "kube-api-access-w56gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.337263 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mndrw\" (UniqueName: \"kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw\") pod \"13c68175-ca4c-4060-ac38-59f953323712\" (UID: \"13c68175-ca4c-4060-ac38-59f953323712\") " Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.337410 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m97k\" (UniqueName: \"kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k\") pod \"635e82b7-cdaf-4cc7-b1a9-6ed165196067\" (UID: \"635e82b7-cdaf-4cc7-b1a9-6ed165196067\") " Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.338129 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w56gb\" (UniqueName: \"kubernetes.io/projected/cfe56a6d-2d84-449b-b5a8-116c5a9891b1-kube-api-access-w56gb\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.346989 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k" (OuterVolumeSpecName: "kube-api-access-2m97k") pod "635e82b7-cdaf-4cc7-b1a9-6ed165196067" (UID: "635e82b7-cdaf-4cc7-b1a9-6ed165196067"). InnerVolumeSpecName "kube-api-access-2m97k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.347685 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw" (OuterVolumeSpecName: "kube-api-access-mndrw") pod "13c68175-ca4c-4060-ac38-59f953323712" (UID: "13c68175-ca4c-4060-ac38-59f953323712"). InnerVolumeSpecName "kube-api-access-mndrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.440126 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mndrw\" (UniqueName: \"kubernetes.io/projected/13c68175-ca4c-4060-ac38-59f953323712-kube-api-access-mndrw\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.440160 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m97k\" (UniqueName: \"kubernetes.io/projected/635e82b7-cdaf-4cc7-b1a9-6ed165196067-kube-api-access-2m97k\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.550442 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerStarted","Data":"55205ef78fd205356193bc8bab3da332b89060f6711f4e3dfbe5cc4a4208d9eb"} Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.550502 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.552625 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7327-account-create-7xldq" event={"ID":"cfe56a6d-2d84-449b-b5a8-116c5a9891b1","Type":"ContainerDied","Data":"9d28969982b15f1930156de42aafec761577ee6e71cede772550721f0c3b7606"} Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.552651 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d28969982b15f1930156de42aafec761577ee6e71cede772550721f0c3b7606" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.552669 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7327-account-create-7xldq" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.554623 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.554669 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9cc9-account-create-pxjr8" event={"ID":"635e82b7-cdaf-4cc7-b1a9-6ed165196067","Type":"ContainerDied","Data":"ea31d90274fac3ae3e4e8b2bbdf85c3f722e0807e8762f9bbfe7f5ed2bbc091a"} Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.554695 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea31d90274fac3ae3e4e8b2bbdf85c3f722e0807e8762f9bbfe7f5ed2bbc091a" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.563607 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-71b5-account-create-v7977" event={"ID":"13c68175-ca4c-4060-ac38-59f953323712","Type":"ContainerDied","Data":"d07f0c082806374cda394c9855359fc387ecc8ec85feb4241255c620c336239f"} Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.563653 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d07f0c082806374cda394c9855359fc387ecc8ec85feb4241255c620c336239f" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.563715 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-71b5-account-create-v7977" Oct 01 16:06:07 crc kubenswrapper[4688]: I1001 16:06:07.594446 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.207165048 podStartE2EDuration="7.594425466s" podCreationTimestamp="2025-10-01 16:06:00 +0000 UTC" firstStartedPulling="2025-10-01 16:06:01.389970307 +0000 UTC m=+1150.740610269" lastFinishedPulling="2025-10-01 16:06:06.777230725 +0000 UTC m=+1156.127870687" observedRunningTime="2025-10-01 16:06:07.58700413 +0000 UTC m=+1156.937644092" watchObservedRunningTime="2025-10-01 16:06:07.594425466 +0000 UTC m=+1156.945065428" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.540150 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.589324 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681026 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w8mn9"] Oct 01 16:06:08 crc kubenswrapper[4688]: E1001 16:06:08.681360 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c68175-ca4c-4060-ac38-59f953323712" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681376 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c68175-ca4c-4060-ac38-59f953323712" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: E1001 16:06:08.681390 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe56a6d-2d84-449b-b5a8-116c5a9891b1" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681396 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe56a6d-2d84-449b-b5a8-116c5a9891b1" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: E1001 16:06:08.681422 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635e82b7-cdaf-4cc7-b1a9-6ed165196067" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681429 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="635e82b7-cdaf-4cc7-b1a9-6ed165196067" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681622 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe56a6d-2d84-449b-b5a8-116c5a9891b1" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681652 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="635e82b7-cdaf-4cc7-b1a9-6ed165196067" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.681663 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c68175-ca4c-4060-ac38-59f953323712" containerName="mariadb-account-create" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.682212 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.686030 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.686703 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t8rcm" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.686857 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.706814 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w8mn9"] Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.764605 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.764682 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.764731 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.764801 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2xpq\" (UniqueName: \"kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.866075 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.866197 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2xpq\" (UniqueName: \"kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.866272 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.866341 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.871246 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.871951 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.881487 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:08 crc kubenswrapper[4688]: I1001 16:06:08.902939 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2xpq\" (UniqueName: \"kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq\") pod \"nova-cell0-conductor-db-sync-w8mn9\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:09 crc kubenswrapper[4688]: I1001 16:06:09.002578 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:09 crc kubenswrapper[4688]: I1001 16:06:09.465973 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w8mn9"] Oct 01 16:06:09 crc kubenswrapper[4688]: I1001 16:06:09.583974 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" event={"ID":"d7aa288b-66a4-423d-8e2c-2f1af84e429a","Type":"ContainerStarted","Data":"b2c3e190dc49f7901065bffed46dfc2b73a2f40e55371afe970473ef8b44401c"} Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.285013 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.286801 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-central-agent" containerID="cri-o://e7c6acb51677d71c306f88abf20ac4f22d30b4bdd928e30314cba5d0988b0be8" gracePeriod=30 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.286895 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="sg-core" containerID="cri-o://3e4014a2d4fafe2104bcb7d74cdbefbffb133cd3dbbdc397f09315b269e9b1f3" gracePeriod=30 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.286959 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-notification-agent" containerID="cri-o://54fc175da73a5d3a8cf1f937e6427e1c6a78aa42ecc69f1e754976144f8ec0ed" gracePeriod=30 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.286944 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="proxy-httpd" containerID="cri-o://55205ef78fd205356193bc8bab3da332b89060f6711f4e3dfbe5cc4a4208d9eb" gracePeriod=30 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.617593 4688 generic.go:334] "Generic (PLEG): container finished" podID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerID="55205ef78fd205356193bc8bab3da332b89060f6711f4e3dfbe5cc4a4208d9eb" exitCode=0 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.617645 4688 generic.go:334] "Generic (PLEG): container finished" podID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerID="3e4014a2d4fafe2104bcb7d74cdbefbffb133cd3dbbdc397f09315b269e9b1f3" exitCode=2 Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.617654 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerDied","Data":"55205ef78fd205356193bc8bab3da332b89060f6711f4e3dfbe5cc4a4208d9eb"} Oct 01 16:06:12 crc kubenswrapper[4688]: I1001 16:06:12.617685 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerDied","Data":"3e4014a2d4fafe2104bcb7d74cdbefbffb133cd3dbbdc397f09315b269e9b1f3"} Oct 01 16:06:13 crc kubenswrapper[4688]: I1001 16:06:13.629537 4688 generic.go:334] "Generic (PLEG): container finished" podID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerID="54fc175da73a5d3a8cf1f937e6427e1c6a78aa42ecc69f1e754976144f8ec0ed" exitCode=0 Oct 01 16:06:13 crc kubenswrapper[4688]: I1001 16:06:13.629815 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerDied","Data":"54fc175da73a5d3a8cf1f937e6427e1c6a78aa42ecc69f1e754976144f8ec0ed"} Oct 01 16:06:15 crc kubenswrapper[4688]: I1001 16:06:15.844495 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="13321c8d-8fe9-4036-9061-0f5466c84b6c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.169:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:06:16 crc kubenswrapper[4688]: I1001 16:06:16.664713 4688 generic.go:334] "Generic (PLEG): container finished" podID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerID="e7c6acb51677d71c306f88abf20ac4f22d30b4bdd928e30314cba5d0988b0be8" exitCode=0 Oct 01 16:06:16 crc kubenswrapper[4688]: I1001 16:06:16.664775 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerDied","Data":"e7c6acb51677d71c306f88abf20ac4f22d30b4bdd928e30314cba5d0988b0be8"} Oct 01 16:06:18 crc kubenswrapper[4688]: I1001 16:06:18.540710 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:06:18 crc kubenswrapper[4688]: I1001 16:06:18.589506 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-555656bf5b-xlntq" podUID="37923e92-dbcc-41a1-8d2f-89d8de59959e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.306334 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407136 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407193 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407210 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407255 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407830 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407954 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.407995 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.408114 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nmsq\" (UniqueName: \"kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq\") pod \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\" (UID: \"16a47361-bb0e-452f-b1f7-d0b5c8c04342\") " Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.408268 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.408904 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.408929 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16a47361-bb0e-452f-b1f7-d0b5c8c04342-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.440751 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq" (OuterVolumeSpecName: "kube-api-access-5nmsq") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "kube-api-access-5nmsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.445640 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts" (OuterVolumeSpecName: "scripts") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.510816 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.511055 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nmsq\" (UniqueName: \"kubernetes.io/projected/16a47361-bb0e-452f-b1f7-d0b5c8c04342-kube-api-access-5nmsq\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.542131 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.580847 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.612398 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.612427 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.643418 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data" (OuterVolumeSpecName: "config-data") pod "16a47361-bb0e-452f-b1f7-d0b5c8c04342" (UID: "16a47361-bb0e-452f-b1f7-d0b5c8c04342"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.706363 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16a47361-bb0e-452f-b1f7-d0b5c8c04342","Type":"ContainerDied","Data":"5160616d9e12c7eac7a0bf15c21c7908ccc5b218dd07e929c848f0ea833fc69c"} Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.706416 4688 scope.go:117] "RemoveContainer" containerID="55205ef78fd205356193bc8bab3da332b89060f6711f4e3dfbe5cc4a4208d9eb" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.706418 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.708377 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" event={"ID":"d7aa288b-66a4-423d-8e2c-2f1af84e429a","Type":"ContainerStarted","Data":"0f04e8d5b0e7873574e92569f8b2f8367aa249018378f57c6b9fbc0f11c565a2"} Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.714451 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16a47361-bb0e-452f-b1f7-d0b5c8c04342-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.741625 4688 scope.go:117] "RemoveContainer" containerID="3e4014a2d4fafe2104bcb7d74cdbefbffb133cd3dbbdc397f09315b269e9b1f3" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.761049 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" podStartSLOduration=2.078794155 podStartE2EDuration="12.761033007s" podCreationTimestamp="2025-10-01 16:06:08 +0000 UTC" firstStartedPulling="2025-10-01 16:06:09.486548316 +0000 UTC m=+1158.837188278" lastFinishedPulling="2025-10-01 16:06:20.168787168 +0000 UTC m=+1169.519427130" observedRunningTime="2025-10-01 16:06:20.732835985 +0000 UTC m=+1170.083475947" watchObservedRunningTime="2025-10-01 16:06:20.761033007 +0000 UTC m=+1170.111672969" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.765887 4688 scope.go:117] "RemoveContainer" containerID="54fc175da73a5d3a8cf1f937e6427e1c6a78aa42ecc69f1e754976144f8ec0ed" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.784358 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.831797 4688 scope.go:117] "RemoveContainer" containerID="e7c6acb51677d71c306f88abf20ac4f22d30b4bdd928e30314cba5d0988b0be8" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.851284 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.871867 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:20 crc kubenswrapper[4688]: E1001 16:06:20.872267 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="sg-core" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872284 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="sg-core" Oct 01 16:06:20 crc kubenswrapper[4688]: E1001 16:06:20.872309 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-notification-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872317 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-notification-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: E1001 16:06:20.872331 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="proxy-httpd" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872338 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="proxy-httpd" Oct 01 16:06:20 crc kubenswrapper[4688]: E1001 16:06:20.872353 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-central-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872360 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-central-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872572 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-central-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872590 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="sg-core" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872600 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="ceilometer-notification-agent" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.872625 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" containerName="proxy-httpd" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.874313 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.877061 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.877249 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.879969 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:20 crc kubenswrapper[4688]: E1001 16:06:20.915284 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16a47361_bb0e_452f_b1f7_d0b5c8c04342.slice/crio-5160616d9e12c7eac7a0bf15c21c7908ccc5b218dd07e929c848f0ea833fc69c\": RecentStats: unable to find data in memory cache]" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945372 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945455 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945568 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxr44\" (UniqueName: \"kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945631 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945787 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945886 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:20 crc kubenswrapper[4688]: I1001 16:06:20.945982 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.047923 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxr44\" (UniqueName: \"kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.047967 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.048035 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.048073 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.048110 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.048238 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.048274 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.049116 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.049150 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.053356 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.053701 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.055222 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.056158 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.071185 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxr44\" (UniqueName: \"kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44\") pod \"ceilometer-0\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.194899 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.195299 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.392636 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16a47361-bb0e-452f-b1f7-d0b5c8c04342" path="/var/lib/kubelet/pods/16a47361-bb0e-452f-b1f7-d0b5c8c04342/volumes" Oct 01 16:06:21 crc kubenswrapper[4688]: I1001 16:06:21.825350 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:22 crc kubenswrapper[4688]: I1001 16:06:22.727512 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerStarted","Data":"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30"} Oct 01 16:06:22 crc kubenswrapper[4688]: I1001 16:06:22.727828 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerStarted","Data":"aba9ef017c4d36730bc5ba38cd566297ec79e39d974c1abe7b1a37036651a3ce"} Oct 01 16:06:23 crc kubenswrapper[4688]: I1001 16:06:23.738268 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerStarted","Data":"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790"} Oct 01 16:06:24 crc kubenswrapper[4688]: I1001 16:06:24.756767 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerStarted","Data":"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66"} Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.770093 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerStarted","Data":"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849"} Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.770690 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-central-agent" containerID="cri-o://63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30" gracePeriod=30 Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.771066 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.771473 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="proxy-httpd" containerID="cri-o://b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849" gracePeriod=30 Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.771604 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="sg-core" containerID="cri-o://faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66" gracePeriod=30 Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.771673 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-notification-agent" containerID="cri-o://864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790" gracePeriod=30 Oct 01 16:06:25 crc kubenswrapper[4688]: I1001 16:06:25.805615 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.226036909 podStartE2EDuration="5.805590999s" podCreationTimestamp="2025-10-01 16:06:20 +0000 UTC" firstStartedPulling="2025-10-01 16:06:21.822166628 +0000 UTC m=+1171.172806590" lastFinishedPulling="2025-10-01 16:06:25.401720718 +0000 UTC m=+1174.752360680" observedRunningTime="2025-10-01 16:06:25.79826499 +0000 UTC m=+1175.148904992" watchObservedRunningTime="2025-10-01 16:06:25.805590999 +0000 UTC m=+1175.156230991" Oct 01 16:06:26 crc kubenswrapper[4688]: I1001 16:06:26.790622 4688 generic.go:334] "Generic (PLEG): container finished" podID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerID="faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66" exitCode=2 Oct 01 16:06:26 crc kubenswrapper[4688]: I1001 16:06:26.790976 4688 generic.go:334] "Generic (PLEG): container finished" podID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerID="864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790" exitCode=0 Oct 01 16:06:26 crc kubenswrapper[4688]: I1001 16:06:26.791039 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerDied","Data":"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66"} Oct 01 16:06:26 crc kubenswrapper[4688]: I1001 16:06:26.791079 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerDied","Data":"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790"} Oct 01 16:06:29 crc kubenswrapper[4688]: I1001 16:06:29.818952 4688 generic.go:334] "Generic (PLEG): container finished" podID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerID="63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30" exitCode=0 Oct 01 16:06:29 crc kubenswrapper[4688]: I1001 16:06:29.819089 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerDied","Data":"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30"} Oct 01 16:06:31 crc kubenswrapper[4688]: I1001 16:06:31.021099 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:06:31 crc kubenswrapper[4688]: I1001 16:06:31.021598 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:06:31 crc kubenswrapper[4688]: I1001 16:06:31.840827 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7aa288b-66a4-423d-8e2c-2f1af84e429a" containerID="0f04e8d5b0e7873574e92569f8b2f8367aa249018378f57c6b9fbc0f11c565a2" exitCode=0 Oct 01 16:06:31 crc kubenswrapper[4688]: I1001 16:06:31.840896 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" event={"ID":"d7aa288b-66a4-423d-8e2c-2f1af84e429a","Type":"ContainerDied","Data":"0f04e8d5b0e7873574e92569f8b2f8367aa249018378f57c6b9fbc0f11c565a2"} Oct 01 16:06:32 crc kubenswrapper[4688]: I1001 16:06:32.707300 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-555656bf5b-xlntq" Oct 01 16:06:32 crc kubenswrapper[4688]: I1001 16:06:32.739259 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:06:32 crc kubenswrapper[4688]: I1001 16:06:32.771986 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:06:32 crc kubenswrapper[4688]: I1001 16:06:32.848931 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon-log" containerID="cri-o://908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc" gracePeriod=30 Oct 01 16:06:32 crc kubenswrapper[4688]: I1001 16:06:32.849251 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" containerID="cri-o://de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4" gracePeriod=30 Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.165887 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.312881 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts\") pod \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.313098 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2xpq\" (UniqueName: \"kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq\") pod \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.313240 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data\") pod \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.313946 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle\") pod \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\" (UID: \"d7aa288b-66a4-423d-8e2c-2f1af84e429a\") " Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.318150 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq" (OuterVolumeSpecName: "kube-api-access-p2xpq") pod "d7aa288b-66a4-423d-8e2c-2f1af84e429a" (UID: "d7aa288b-66a4-423d-8e2c-2f1af84e429a"). InnerVolumeSpecName "kube-api-access-p2xpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.323717 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts" (OuterVolumeSpecName: "scripts") pod "d7aa288b-66a4-423d-8e2c-2f1af84e429a" (UID: "d7aa288b-66a4-423d-8e2c-2f1af84e429a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.370795 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7aa288b-66a4-423d-8e2c-2f1af84e429a" (UID: "d7aa288b-66a4-423d-8e2c-2f1af84e429a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.376193 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data" (OuterVolumeSpecName: "config-data") pod "d7aa288b-66a4-423d-8e2c-2f1af84e429a" (UID: "d7aa288b-66a4-423d-8e2c-2f1af84e429a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.428058 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.428087 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.428161 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7aa288b-66a4-423d-8e2c-2f1af84e429a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.428403 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2xpq\" (UniqueName: \"kubernetes.io/projected/d7aa288b-66a4-423d-8e2c-2f1af84e429a-kube-api-access-p2xpq\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.859077 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" event={"ID":"d7aa288b-66a4-423d-8e2c-2f1af84e429a","Type":"ContainerDied","Data":"b2c3e190dc49f7901065bffed46dfc2b73a2f40e55371afe970473ef8b44401c"} Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.859116 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2c3e190dc49f7901065bffed46dfc2b73a2f40e55371afe970473ef8b44401c" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.859140 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w8mn9" Oct 01 16:06:33 crc kubenswrapper[4688]: I1001 16:06:33.977102 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:06:34 crc kubenswrapper[4688]: E1001 16:06:33.978511 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7aa288b-66a4-423d-8e2c-2f1af84e429a" containerName="nova-cell0-conductor-db-sync" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.978557 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7aa288b-66a4-423d-8e2c-2f1af84e429a" containerName="nova-cell0-conductor-db-sync" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.978833 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7aa288b-66a4-423d-8e2c-2f1af84e429a" containerName="nova-cell0-conductor-db-sync" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.979602 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.987067 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.987627 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t8rcm" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:33.997630 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.039034 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.039119 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.039162 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqc59\" (UniqueName: \"kubernetes.io/projected/f64f51ee-f2ae-4c33-babf-67838b0f0135-kube-api-access-kqc59\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.141312 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.141734 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqc59\" (UniqueName: \"kubernetes.io/projected/f64f51ee-f2ae-4c33-babf-67838b0f0135-kube-api-access-kqc59\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.141844 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.147620 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.148592 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64f51ee-f2ae-4c33-babf-67838b0f0135-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.160036 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqc59\" (UniqueName: \"kubernetes.io/projected/f64f51ee-f2ae-4c33-babf-67838b0f0135-kube-api-access-kqc59\") pod \"nova-cell0-conductor-0\" (UID: \"f64f51ee-f2ae-4c33-babf-67838b0f0135\") " pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.307915 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.843043 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 16:06:34 crc kubenswrapper[4688]: I1001 16:06:34.870463 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f64f51ee-f2ae-4c33-babf-67838b0f0135","Type":"ContainerStarted","Data":"1d0502f28de0ee973c8ad868765b47a24481f27ae582ba06c8dc0cb9cde99784"} Oct 01 16:06:35 crc kubenswrapper[4688]: I1001 16:06:35.887601 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f64f51ee-f2ae-4c33-babf-67838b0f0135","Type":"ContainerStarted","Data":"7c692ea5f9a4722eaace712ae97ac35c05edd58055187dd791887870aa7aa770"} Oct 01 16:06:35 crc kubenswrapper[4688]: I1001 16:06:35.889335 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:35 crc kubenswrapper[4688]: I1001 16:06:35.937981 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.937954923 podStartE2EDuration="2.937954923s" podCreationTimestamp="2025-10-01 16:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:35.922687589 +0000 UTC m=+1185.273327581" watchObservedRunningTime="2025-10-01 16:06:35.937954923 +0000 UTC m=+1185.288594895" Oct 01 16:06:36 crc kubenswrapper[4688]: I1001 16:06:36.899944 4688 generic.go:334] "Generic (PLEG): container finished" podID="85db256a-98d5-4bca-82ac-098d49b243e8" containerID="de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4" exitCode=0 Oct 01 16:06:36 crc kubenswrapper[4688]: I1001 16:06:36.900018 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerDied","Data":"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4"} Oct 01 16:06:36 crc kubenswrapper[4688]: I1001 16:06:36.901232 4688 scope.go:117] "RemoveContainer" containerID="db1e7500239d702cf0eb165c9c134f6170374d4a8e1d9485103f973d086581f2" Oct 01 16:06:38 crc kubenswrapper[4688]: I1001 16:06:38.538682 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.359131 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.928890 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ds7rw"] Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.934996 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.938221 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.938223 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 01 16:06:44 crc kubenswrapper[4688]: I1001 16:06:44.949282 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ds7rw"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.065283 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.065353 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2b5f\" (UniqueName: \"kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.065392 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.065444 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.154990 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.156624 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.161001 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.166540 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.166608 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2b5f\" (UniqueName: \"kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.166649 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.166698 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.175415 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.180997 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.183628 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.185798 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.188716 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.193106 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.203577 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.222021 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2b5f\" (UniqueName: \"kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f\") pod \"nova-cell0-cell-mapping-ds7rw\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.244657 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.266148 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267714 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267764 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267804 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267821 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7w6t\" (UniqueName: \"kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267853 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjzb7\" (UniqueName: \"kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267910 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.267945 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369511 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369584 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369605 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7w6t\" (UniqueName: \"kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369643 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjzb7\" (UniqueName: \"kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369700 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369735 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.369779 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.373591 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.374808 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.377029 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.379844 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.394436 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.398775 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjzb7\" (UniqueName: \"kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.399997 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.420382 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.424547 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.424744 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.426186 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.437941 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7w6t\" (UniqueName: \"kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t\") pod \"nova-scheduler-0\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.449636 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.471796 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data\") pod \"nova-metadata-0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.472794 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.472820 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.472924 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svrvd\" (UniqueName: \"kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.511374 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574034 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8jpj\" (UniqueName: \"kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574156 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574205 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574261 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574282 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574309 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574359 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.574396 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svrvd\" (UniqueName: \"kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.578467 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.580031 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.580442 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.584750 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.586851 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.598179 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.612954 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.615455 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svrvd\" (UniqueName: \"kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679306 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679341 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679394 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679424 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679517 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plp9b\" (UniqueName: \"kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679568 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679649 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8jpj\" (UniqueName: \"kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679684 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679721 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.679751 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.686546 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.686662 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.686919 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.687074 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.687122 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.710792 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8jpj\" (UniqueName: \"kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj\") pod \"dnsmasq-dns-bccf8f775-dj9ns\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.780917 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.780987 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.781020 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plp9b\" (UniqueName: \"kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.781042 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.781502 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.787577 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.793064 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.818064 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plp9b\" (UniqueName: \"kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b\") pod \"nova-api-0\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.836948 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.881651 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.932005 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:06:45 crc kubenswrapper[4688]: I1001 16:06:45.996876 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ds7rw"] Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.266588 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.286962 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:06:46 crc kubenswrapper[4688]: W1001 16:06:46.300900 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod469801bc_76a8_4241_a4e1_7fbd9bd9d39f.slice/crio-2c5e65574d2b76dddf9dd1b9e6b11229e59e45fdb2f3c3b32d0a3369e1e9f1f5 WatchSource:0}: Error finding container 2c5e65574d2b76dddf9dd1b9e6b11229e59e45fdb2f3c3b32d0a3369e1e9f1f5: Status 404 returned error can't find the container with id 2c5e65574d2b76dddf9dd1b9e6b11229e59e45fdb2f3c3b32d0a3369e1e9f1f5 Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.446487 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5xckd"] Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.448890 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.453638 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.453813 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.459151 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5xckd"] Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.509013 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.509182 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.509297 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.509324 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7g67\" (UniqueName: \"kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.557736 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.611160 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.611489 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7g67\" (UniqueName: \"kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.612299 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.623042 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.624739 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.615296 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.626569 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.632238 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7g67\" (UniqueName: \"kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67\") pod \"nova-cell1-conductor-db-sync-5xckd\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.775897 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:06:46 crc kubenswrapper[4688]: W1001 16:06:46.776868 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51006f0d_a088_4424_be58_fb677966614d.slice/crio-39d103a06934eaab47bd10275909b8966be4870e071bdc38bd642a99315018e7 WatchSource:0}: Error finding container 39d103a06934eaab47bd10275909b8966be4870e071bdc38bd642a99315018e7: Status 404 returned error can't find the container with id 39d103a06934eaab47bd10275909b8966be4870e071bdc38bd642a99315018e7 Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.801720 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:46 crc kubenswrapper[4688]: W1001 16:06:46.845796 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e75ac4_87bc_431a_9ec5_943ec836a9bc.slice/crio-0ce4f3165613fb17645072ad70d7d66415d5c155fe33255b9a1d3dafb0999257 WatchSource:0}: Error finding container 0ce4f3165613fb17645072ad70d7d66415d5c155fe33255b9a1d3dafb0999257: Status 404 returned error can't find the container with id 0ce4f3165613fb17645072ad70d7d66415d5c155fe33255b9a1d3dafb0999257 Oct 01 16:06:46 crc kubenswrapper[4688]: I1001 16:06:46.846900 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.095267 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerStarted","Data":"10a3e08635ae5784c03e3e764a5865f3de75608ef8d3bf24dae6097e3320e2e0"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.103111 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1682b567-2a60-4625-a0cd-d44627b2f7ec","Type":"ContainerStarted","Data":"e05635b5fb5965e985007b6b052796d1e21468ccb090d5a21c69bfa6a144b61f"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.119423 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ds7rw" event={"ID":"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb","Type":"ContainerStarted","Data":"39a07c3ed94706947f921a9d90921b151986bc3d04e5a090e52081b6183353f8"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.119476 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ds7rw" event={"ID":"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb","Type":"ContainerStarted","Data":"ece2388bf72162ee509c5eb0acec6b8facf11645321a6b6d75939402c117117a"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.132037 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"469801bc-76a8-4241-a4e1-7fbd9bd9d39f","Type":"ContainerStarted","Data":"2c5e65574d2b76dddf9dd1b9e6b11229e59e45fdb2f3c3b32d0a3369e1e9f1f5"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.148003 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ds7rw" podStartSLOduration=3.147985829 podStartE2EDuration="3.147985829s" podCreationTimestamp="2025-10-01 16:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:47.147764562 +0000 UTC m=+1196.498404524" watchObservedRunningTime="2025-10-01 16:06:47.147985829 +0000 UTC m=+1196.498625791" Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.149022 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" event={"ID":"51006f0d-a088-4424-be58-fb677966614d","Type":"ContainerStarted","Data":"39d103a06934eaab47bd10275909b8966be4870e071bdc38bd642a99315018e7"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.157617 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerStarted","Data":"0ce4f3165613fb17645072ad70d7d66415d5c155fe33255b9a1d3dafb0999257"} Oct 01 16:06:47 crc kubenswrapper[4688]: I1001 16:06:47.341456 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5xckd"] Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.182246 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5xckd" event={"ID":"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1","Type":"ContainerStarted","Data":"f9087ea8ce0a897ec369432bf53dcca1b4b86a133e060a4d45f85249cb82554f"} Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.182592 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5xckd" event={"ID":"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1","Type":"ContainerStarted","Data":"e0fd4a5faf646f2473236731754d1e0b6a26ef29e411ac79734392a4bb2e25cb"} Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.188010 4688 generic.go:334] "Generic (PLEG): container finished" podID="51006f0d-a088-4424-be58-fb677966614d" containerID="07d2c36542a2d4d8109f9b3985d9399d72d8d66a2a2551cba27a4a6f4bb5d693" exitCode=0 Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.188686 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" event={"ID":"51006f0d-a088-4424-be58-fb677966614d","Type":"ContainerDied","Data":"07d2c36542a2d4d8109f9b3985d9399d72d8d66a2a2551cba27a4a6f4bb5d693"} Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.205614 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5xckd" podStartSLOduration=2.205600849 podStartE2EDuration="2.205600849s" podCreationTimestamp="2025-10-01 16:06:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:48.196318015 +0000 UTC m=+1197.546957977" watchObservedRunningTime="2025-10-01 16:06:48.205600849 +0000 UTC m=+1197.556240811" Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.540115 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.956533 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:06:48 crc kubenswrapper[4688]: I1001 16:06:48.979511 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.227189 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.230741 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1682b567-2a60-4625-a0cd-d44627b2f7ec","Type":"ContainerStarted","Data":"f87044b84436f2a688c3f949cb91e14f043f3c4c85739aade57d998dac4b2ff4"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.230775 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1682b567-2a60-4625-a0cd-d44627b2f7ec" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f87044b84436f2a688c3f949cb91e14f043f3c4c85739aade57d998dac4b2ff4" gracePeriod=30 Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.233501 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"469801bc-76a8-4241-a4e1-7fbd9bd9d39f","Type":"ContainerStarted","Data":"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.238383 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" event={"ID":"51006f0d-a088-4424-be58-fb677966614d","Type":"ContainerStarted","Data":"94373c3c90ca14b126a9b2291e350ba725ec647d2250b0255943597ed46e07ca"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.238425 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.243983 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerStarted","Data":"0b89eaafca7a994efabf5e1551a31ac25d4fe24c8a2a6dc13ed90b43ec384afd"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.244019 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerStarted","Data":"3cf9a7790f2460c71b210d0810788a3901e37b6d665e94a08fc15561b399539c"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.252640 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerStarted","Data":"58f1884793ca8c394b4449d587709fe97e059bdbb69059104602956929edf013"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.252778 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerStarted","Data":"f630d5f9a142198bb5b70976cb8d5c6e450a2cc7c8451370467ffd9e3ad2d12c"} Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.253081 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-log" containerID="cri-o://f630d5f9a142198bb5b70976cb8d5c6e450a2cc7c8451370467ffd9e3ad2d12c" gracePeriod=30 Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.254368 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-metadata" containerID="cri-o://58f1884793ca8c394b4449d587709fe97e059bdbb69059104602956929edf013" gracePeriod=30 Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.258692 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.133306671 podStartE2EDuration="6.258677262s" podCreationTimestamp="2025-10-01 16:06:45 +0000 UTC" firstStartedPulling="2025-10-01 16:06:46.547812053 +0000 UTC m=+1195.898452015" lastFinishedPulling="2025-10-01 16:06:50.673182634 +0000 UTC m=+1200.023822606" observedRunningTime="2025-10-01 16:06:51.249885692 +0000 UTC m=+1200.600525654" watchObservedRunningTime="2025-10-01 16:06:51.258677262 +0000 UTC m=+1200.609317224" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.269869 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" podStartSLOduration=6.26985062 podStartE2EDuration="6.26985062s" podCreationTimestamp="2025-10-01 16:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:06:51.264981231 +0000 UTC m=+1200.615621183" watchObservedRunningTime="2025-10-01 16:06:51.26985062 +0000 UTC m=+1200.620490582" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.293993 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.503418721 podStartE2EDuration="6.293976846s" podCreationTimestamp="2025-10-01 16:06:45 +0000 UTC" firstStartedPulling="2025-10-01 16:06:46.863480144 +0000 UTC m=+1196.214120106" lastFinishedPulling="2025-10-01 16:06:50.654038239 +0000 UTC m=+1200.004678231" observedRunningTime="2025-10-01 16:06:51.286202735 +0000 UTC m=+1200.636842707" watchObservedRunningTime="2025-10-01 16:06:51.293976846 +0000 UTC m=+1200.644616808" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.305471 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.980683888 podStartE2EDuration="6.305452292s" podCreationTimestamp="2025-10-01 16:06:45 +0000 UTC" firstStartedPulling="2025-10-01 16:06:46.307249659 +0000 UTC m=+1195.657889621" lastFinishedPulling="2025-10-01 16:06:50.632018063 +0000 UTC m=+1199.982658025" observedRunningTime="2025-10-01 16:06:51.298936337 +0000 UTC m=+1200.649576299" watchObservedRunningTime="2025-10-01 16:06:51.305452292 +0000 UTC m=+1200.656092254" Oct 01 16:06:51 crc kubenswrapper[4688]: I1001 16:06:51.326579 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.931690844 podStartE2EDuration="6.326560193s" podCreationTimestamp="2025-10-01 16:06:45 +0000 UTC" firstStartedPulling="2025-10-01 16:06:46.313103845 +0000 UTC m=+1195.663743807" lastFinishedPulling="2025-10-01 16:06:50.707973194 +0000 UTC m=+1200.058613156" observedRunningTime="2025-10-01 16:06:51.316149377 +0000 UTC m=+1200.666789329" watchObservedRunningTime="2025-10-01 16:06:51.326560193 +0000 UTC m=+1200.677200145" Oct 01 16:06:52 crc kubenswrapper[4688]: I1001 16:06:52.265593 4688 generic.go:334] "Generic (PLEG): container finished" podID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerID="f630d5f9a142198bb5b70976cb8d5c6e450a2cc7c8451370467ffd9e3ad2d12c" exitCode=143 Oct 01 16:06:52 crc kubenswrapper[4688]: I1001 16:06:52.265674 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerDied","Data":"f630d5f9a142198bb5b70976cb8d5c6e450a2cc7c8451370467ffd9e3ad2d12c"} Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.300729 4688 generic.go:334] "Generic (PLEG): container finished" podID="131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" containerID="39a07c3ed94706947f921a9d90921b151986bc3d04e5a090e52081b6183353f8" exitCode=0 Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.300786 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ds7rw" event={"ID":"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb","Type":"ContainerDied","Data":"39a07c3ed94706947f921a9d90921b151986bc3d04e5a090e52081b6183353f8"} Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.512567 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.512616 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.598795 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.599936 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.623183 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.803079 4688 scope.go:117] "RemoveContainer" containerID="49f49eb4fb0cff420e54ff2ee89dd5a7bf8b1456b10e2ee6179ca580ba7b3e6e" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.838001 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.885987 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.949332 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:06:55 crc kubenswrapper[4688]: I1001 16:06:55.949404 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.017554 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.030963 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="dnsmasq-dns" containerID="cri-o://b5e01ee8496aba3e1f6df99b638140f61b7dcae71b6229a1d9f47f172dd56e97" gracePeriod=10 Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.040047 4688 scope.go:117] "RemoveContainer" containerID="7474f86f53785dab4b5ec9805e0aebc8a634792ff318211cee2c64bc1e13bb19" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.154858 4688 scope.go:117] "RemoveContainer" containerID="b63d98fc79af68824e8387a765c0e2b65d961f38b134ae5dccb598541ea7d7d7" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.276087 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.337601 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.337712 4688 generic.go:334] "Generic (PLEG): container finished" podID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerID="b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849" exitCode=137 Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.337809 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerDied","Data":"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849"} Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.337836 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76a8f837-f152-4b08-ab79-b3f6a620127f","Type":"ContainerDied","Data":"aba9ef017c4d36730bc5ba38cd566297ec79e39d974c1abe7b1a37036651a3ce"} Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.337852 4688 scope.go:117] "RemoveContainer" containerID="b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.342895 4688 generic.go:334] "Generic (PLEG): container finished" podID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerID="b5e01ee8496aba3e1f6df99b638140f61b7dcae71b6229a1d9f47f172dd56e97" exitCode=0 Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.342945 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerDied","Data":"b5e01ee8496aba3e1f6df99b638140f61b7dcae71b6229a1d9f47f172dd56e97"} Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348645 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348704 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348739 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348760 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxr44\" (UniqueName: \"kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348786 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348827 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.348868 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml\") pod \"76a8f837-f152-4b08-ab79-b3f6a620127f\" (UID: \"76a8f837-f152-4b08-ab79-b3f6a620127f\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.349915 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.349990 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.350548 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.350561 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76a8f837-f152-4b08-ab79-b3f6a620127f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.360704 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44" (OuterVolumeSpecName: "kube-api-access-xxr44") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "kube-api-access-xxr44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.365912 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts" (OuterVolumeSpecName: "scripts") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.400100 4688 scope.go:117] "RemoveContainer" containerID="faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.407385 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.436776 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.451646 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxr44\" (UniqueName: \"kubernetes.io/projected/76a8f837-f152-4b08-ab79-b3f6a620127f-kube-api-access-xxr44\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.452833 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.452924 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.486613 4688 scope.go:117] "RemoveContainer" containerID="864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.526667 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.547839 4688 scope.go:117] "RemoveContainer" containerID="63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.555498 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.575373 4688 scope.go:117] "RemoveContainer" containerID="b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.575990 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849\": container with ID starting with b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849 not found: ID does not exist" containerID="b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576025 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849"} err="failed to get container status \"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849\": rpc error: code = NotFound desc = could not find container \"b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849\": container with ID starting with b1b6355ad79a27f354db13e3785e83b8af6d4dde454f07180c36f6ba83573849 not found: ID does not exist" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576045 4688 scope.go:117] "RemoveContainer" containerID="faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.576236 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66\": container with ID starting with faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66 not found: ID does not exist" containerID="faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576267 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66"} err="failed to get container status \"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66\": rpc error: code = NotFound desc = could not find container \"faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66\": container with ID starting with faec177e90254be8051906d2a05b03daa3eafbb47ed71f6a9145521c8854fc66 not found: ID does not exist" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576280 4688 scope.go:117] "RemoveContainer" containerID="864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.576471 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790\": container with ID starting with 864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790 not found: ID does not exist" containerID="864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576490 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790"} err="failed to get container status \"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790\": rpc error: code = NotFound desc = could not find container \"864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790\": container with ID starting with 864045b9fbf4c06cd32eb9c62d7ed8f63a2a6f807a8e9b54aa60c3f59641a790 not found: ID does not exist" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576512 4688 scope.go:117] "RemoveContainer" containerID="63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.576776 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30\": container with ID starting with 63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30 not found: ID does not exist" containerID="63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.576796 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30"} err="failed to get container status \"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30\": rpc error: code = NotFound desc = could not find container \"63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30\": container with ID starting with 63e58178e3c282b396c43e37a33c3c141adc9c5d5db16f5cb1a2567d7d2d4a30 not found: ID does not exist" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.579248 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data" (OuterVolumeSpecName: "config-data") pod "76a8f837-f152-4b08-ab79-b3f6a620127f" (UID: "76a8f837-f152-4b08-ab79-b3f6a620127f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.596848 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.656956 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a8f837-f152-4b08-ab79-b3f6a620127f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.735661 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.751150 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.757882 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.758000 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjv75\" (UniqueName: \"kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.758053 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.758076 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.758100 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.758141 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc\") pod \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\" (UID: \"7ff4e6f3-7434-463c-9701-417d73f0a4bb\") " Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.783165 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787319 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-central-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787350 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-central-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787414 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="sg-core" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787424 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="sg-core" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787460 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="dnsmasq-dns" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787469 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="dnsmasq-dns" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787490 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-notification-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787498 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-notification-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787511 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="proxy-httpd" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787563 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="proxy-httpd" Oct 01 16:06:56 crc kubenswrapper[4688]: E1001 16:06:56.787582 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="init" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787599 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="init" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787942 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" containerName="dnsmasq-dns" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787978 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-notification-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.787996 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="proxy-httpd" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.788005 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="sg-core" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.788026 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" containerName="ceilometer-central-agent" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.792009 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.793392 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75" (OuterVolumeSpecName: "kube-api-access-qjv75") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "kube-api-access-qjv75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.799091 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.801146 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.804349 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.860601 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjv75\" (UniqueName: \"kubernetes.io/projected/7ff4e6f3-7434-463c-9701-417d73f0a4bb-kube-api-access-qjv75\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.939853 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.961819 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962074 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962218 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962344 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962450 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mngg\" (UniqueName: \"kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962572 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962685 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.962872 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.970199 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.971004 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:06:56 crc kubenswrapper[4688]: I1001 16:06:56.975693 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:56.997671 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config" (OuterVolumeSpecName: "config") pod "7ff4e6f3-7434-463c-9701-417d73f0a4bb" (UID: "7ff4e6f3-7434-463c-9701-417d73f0a4bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.032482 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.032583 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064444 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064502 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064560 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064624 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064646 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mngg\" (UniqueName: \"kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064691 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064723 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064836 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064846 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064854 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.064863 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ff4e6f3-7434-463c-9701-417d73f0a4bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.065931 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.066210 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.074653 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.075209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.084225 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.085035 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.089641 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mngg\" (UniqueName: \"kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg\") pod \"ceilometer-0\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.168631 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.269458 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data\") pod \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.269572 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle\") pod \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.269651 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2b5f\" (UniqueName: \"kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f\") pod \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.269765 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts\") pod \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\" (UID: \"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb\") " Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.282994 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts" (OuterVolumeSpecName: "scripts") pod "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" (UID: "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.284006 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f" (OuterVolumeSpecName: "kube-api-access-v2b5f") pod "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" (UID: "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb"). InnerVolumeSpecName "kube-api-access-v2b5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.308564 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data" (OuterVolumeSpecName: "config-data") pod "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" (UID: "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.324192 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" (UID: "131a8df6-8ea9-48d0-a8e3-db8bec03dbcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.359919 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" event={"ID":"7ff4e6f3-7434-463c-9701-417d73f0a4bb","Type":"ContainerDied","Data":"eeacaf2a627648e48914abefd726c9439a746815f713919c5089dc5328e3a30a"} Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.359988 4688 scope.go:117] "RemoveContainer" containerID="b5e01ee8496aba3e1f6df99b638140f61b7dcae71b6229a1d9f47f172dd56e97" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.360118 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q5p5z" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.370137 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ds7rw" event={"ID":"131a8df6-8ea9-48d0-a8e3-db8bec03dbcb","Type":"ContainerDied","Data":"ece2388bf72162ee509c5eb0acec6b8facf11645321a6b6d75939402c117117a"} Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.370179 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ece2388bf72162ee509c5eb0acec6b8facf11645321a6b6d75939402c117117a" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.370242 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ds7rw" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.373624 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.373654 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.373666 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2b5f\" (UniqueName: \"kubernetes.io/projected/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-kube-api-access-v2b5f\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.373677 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.375001 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.375445 4688 generic.go:334] "Generic (PLEG): container finished" podID="3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" containerID="f9087ea8ce0a897ec369432bf53dcca1b4b86a133e060a4d45f85249cb82554f" exitCode=0 Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.375568 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5xckd" event={"ID":"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1","Type":"ContainerDied","Data":"f9087ea8ce0a897ec369432bf53dcca1b4b86a133e060a4d45f85249cb82554f"} Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.404430 4688 scope.go:117] "RemoveContainer" containerID="ae5a5f263d62c9c6df8c85192db68e99dc40a7f87a8d652e58685339c0e4fb66" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.413842 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a8f837-f152-4b08-ab79-b3f6a620127f" path="/var/lib/kubelet/pods/76a8f837-f152-4b08-ab79-b3f6a620127f/volumes" Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.437632 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.442585 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q5p5z"] Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.518498 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.518701 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-log" containerID="cri-o://3cf9a7790f2460c71b210d0810788a3901e37b6d665e94a08fc15561b399539c" gracePeriod=30 Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.519040 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-api" containerID="cri-o://0b89eaafca7a994efabf5e1551a31ac25d4fe24c8a2a6dc13ed90b43ec384afd" gracePeriod=30 Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.585270 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:06:57 crc kubenswrapper[4688]: I1001 16:06:57.939004 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:06:57 crc kubenswrapper[4688]: W1001 16:06:57.945555 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a9f48e4_a0e0_4227_b108_8caea9529999.slice/crio-c9cca9d35829cf7ea995f3c1e50ac891396d697441aaddbda3f1b4867037e90c WatchSource:0}: Error finding container c9cca9d35829cf7ea995f3c1e50ac891396d697441aaddbda3f1b4867037e90c: Status 404 returned error can't find the container with id c9cca9d35829cf7ea995f3c1e50ac891396d697441aaddbda3f1b4867037e90c Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.394774 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerStarted","Data":"c9cca9d35829cf7ea995f3c1e50ac891396d697441aaddbda3f1b4867037e90c"} Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.409553 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerID="3cf9a7790f2460c71b210d0810788a3901e37b6d665e94a08fc15561b399539c" exitCode=143 Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.409707 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerDied","Data":"3cf9a7790f2460c71b210d0810788a3901e37b6d665e94a08fc15561b399539c"} Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.542726 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-654b94b7dd-g2k9q" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.543022 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.723265 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.906694 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data\") pod \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.906904 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7g67\" (UniqueName: \"kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67\") pod \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.906947 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts\") pod \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.906982 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle\") pod \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\" (UID: \"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1\") " Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.929721 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts" (OuterVolumeSpecName: "scripts") pod "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" (UID: "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.941749 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67" (OuterVolumeSpecName: "kube-api-access-t7g67") pod "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" (UID: "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1"). InnerVolumeSpecName "kube-api-access-t7g67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.952819 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data" (OuterVolumeSpecName: "config-data") pod "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" (UID: "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:58 crc kubenswrapper[4688]: I1001 16:06:58.975694 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" (UID: "3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.010619 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.010649 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7g67\" (UniqueName: \"kubernetes.io/projected/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-kube-api-access-t7g67\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.010660 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.010669 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.394604 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff4e6f3-7434-463c-9701-417d73f0a4bb" path="/var/lib/kubelet/pods/7ff4e6f3-7434-463c-9701-417d73f0a4bb/volumes" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.419343 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerStarted","Data":"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32"} Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.419401 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerStarted","Data":"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd"} Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.423136 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerName="nova-scheduler-scheduler" containerID="cri-o://99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" gracePeriod=30 Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.423567 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5xckd" event={"ID":"3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1","Type":"ContainerDied","Data":"e0fd4a5faf646f2473236731754d1e0b6a26ef29e411ac79734392a4bb2e25cb"} Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.423653 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0fd4a5faf646f2473236731754d1e0b6a26ef29e411ac79734392a4bb2e25cb" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.427265 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5xckd" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.474997 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:06:59 crc kubenswrapper[4688]: E1001 16:06:59.475554 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" containerName="nova-cell1-conductor-db-sync" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.475568 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" containerName="nova-cell1-conductor-db-sync" Oct 01 16:06:59 crc kubenswrapper[4688]: E1001 16:06:59.475617 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" containerName="nova-manage" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.475624 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" containerName="nova-manage" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.475854 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" containerName="nova-cell1-conductor-db-sync" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.475878 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" containerName="nova-manage" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.476647 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.486157 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.494928 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.639598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvbzq\" (UniqueName: \"kubernetes.io/projected/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-kube-api-access-qvbzq\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.639673 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.639693 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.741284 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.741329 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.741458 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvbzq\" (UniqueName: \"kubernetes.io/projected/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-kube-api-access-qvbzq\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.747011 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.754108 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.761252 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvbzq\" (UniqueName: \"kubernetes.io/projected/72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2-kube-api-access-qvbzq\") pod \"nova-cell1-conductor-0\" (UID: \"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 16:06:59 crc kubenswrapper[4688]: I1001 16:06:59.814963 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 16:07:00 crc kubenswrapper[4688]: I1001 16:07:00.381951 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 16:07:00 crc kubenswrapper[4688]: I1001 16:07:00.436623 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2","Type":"ContainerStarted","Data":"dbbf7725ed5609d6341e1b0f07cf5444fe9783055727ae3308683b5f5abbceaf"} Oct 01 16:07:00 crc kubenswrapper[4688]: I1001 16:07:00.440438 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerStarted","Data":"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1"} Oct 01 16:07:00 crc kubenswrapper[4688]: E1001 16:07:00.603862 4688 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:07:00 crc kubenswrapper[4688]: E1001 16:07:00.605418 4688 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:07:00 crc kubenswrapper[4688]: E1001 16:07:00.606508 4688 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 16:07:00 crc kubenswrapper[4688]: E1001 16:07:00.606549 4688 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerName="nova-scheduler-scheduler" Oct 01 16:07:01 crc kubenswrapper[4688]: I1001 16:07:01.453361 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2","Type":"ContainerStarted","Data":"498a49c27fe687dbe753812c1fca0854286c6068722361faa3c5fb2b69743209"} Oct 01 16:07:01 crc kubenswrapper[4688]: I1001 16:07:01.453581 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 16:07:01 crc kubenswrapper[4688]: I1001 16:07:01.485517 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.485498783 podStartE2EDuration="2.485498783s" podCreationTimestamp="2025-10-01 16:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:01.472019649 +0000 UTC m=+1210.822659611" watchObservedRunningTime="2025-10-01 16:07:01.485498783 +0000 UTC m=+1210.836138745" Oct 01 16:07:02 crc kubenswrapper[4688]: I1001 16:07:02.478311 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerStarted","Data":"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b"} Oct 01 16:07:02 crc kubenswrapper[4688]: I1001 16:07:02.480267 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:07:02 crc kubenswrapper[4688]: I1001 16:07:02.512356 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.680741836 podStartE2EDuration="6.512333817s" podCreationTimestamp="2025-10-01 16:06:56 +0000 UTC" firstStartedPulling="2025-10-01 16:06:57.948257966 +0000 UTC m=+1207.298897928" lastFinishedPulling="2025-10-01 16:07:01.779849947 +0000 UTC m=+1211.130489909" observedRunningTime="2025-10-01 16:07:02.50328788 +0000 UTC m=+1211.853927862" watchObservedRunningTime="2025-10-01 16:07:02.512333817 +0000 UTC m=+1211.862973789" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.231692 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.320669 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle\") pod \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.321282 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7w6t\" (UniqueName: \"kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t\") pod \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.322435 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data\") pod \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\" (UID: \"469801bc-76a8-4241-a4e1-7fbd9bd9d39f\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.337692 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t" (OuterVolumeSpecName: "kube-api-access-g7w6t") pod "469801bc-76a8-4241-a4e1-7fbd9bd9d39f" (UID: "469801bc-76a8-4241-a4e1-7fbd9bd9d39f"). InnerVolumeSpecName "kube-api-access-g7w6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.361801 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data" (OuterVolumeSpecName: "config-data") pod "469801bc-76a8-4241-a4e1-7fbd9bd9d39f" (UID: "469801bc-76a8-4241-a4e1-7fbd9bd9d39f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.388616 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "469801bc-76a8-4241-a4e1-7fbd9bd9d39f" (UID: "469801bc-76a8-4241-a4e1-7fbd9bd9d39f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.427321 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7w6t\" (UniqueName: \"kubernetes.io/projected/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-kube-api-access-g7w6t\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.427349 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.427360 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469801bc-76a8-4241-a4e1-7fbd9bd9d39f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.431193 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.498018 4688 generic.go:334] "Generic (PLEG): container finished" podID="85db256a-98d5-4bca-82ac-098d49b243e8" containerID="908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc" exitCode=137 Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.498093 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerDied","Data":"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc"} Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.498120 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-654b94b7dd-g2k9q" event={"ID":"85db256a-98d5-4bca-82ac-098d49b243e8","Type":"ContainerDied","Data":"0b91ae9614216b0bfe30f98a0e38341b10439941cfab805630039fa06af1b15e"} Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.498135 4688 scope.go:117] "RemoveContainer" containerID="de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.498241 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-654b94b7dd-g2k9q" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.504592 4688 generic.go:334] "Generic (PLEG): container finished" podID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" exitCode=0 Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.504645 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"469801bc-76a8-4241-a4e1-7fbd9bd9d39f","Type":"ContainerDied","Data":"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560"} Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.504668 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"469801bc-76a8-4241-a4e1-7fbd9bd9d39f","Type":"ContainerDied","Data":"2c5e65574d2b76dddf9dd1b9e6b11229e59e45fdb2f3c3b32d0a3369e1e9f1f5"} Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.504710 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.520091 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerID="0b89eaafca7a994efabf5e1551a31ac25d4fe24c8a2a6dc13ed90b43ec384afd" exitCode=0 Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.520165 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerDied","Data":"0b89eaafca7a994efabf5e1551a31ac25d4fe24c8a2a6dc13ed90b43ec384afd"} Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.533554 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.562853 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.576422 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.577163 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon-log" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.577246 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon-log" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.577334 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.577410 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.577517 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerName="nova-scheduler-scheduler" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.577628 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerName="nova-scheduler-scheduler" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.577719 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.577787 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.578057 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.578170 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon-log" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.578249 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" containerName="nova-scheduler-scheduler" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.579137 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.584345 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.610862 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631053 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631100 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631177 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631203 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631243 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631291 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.631424 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwnr\" (UniqueName: \"kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr\") pod \"85db256a-98d5-4bca-82ac-098d49b243e8\" (UID: \"85db256a-98d5-4bca-82ac-098d49b243e8\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.632345 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs" (OuterVolumeSpecName: "logs") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.633169 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85db256a-98d5-4bca-82ac-098d49b243e8-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.641810 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr" (OuterVolumeSpecName: "kube-api-access-wpwnr") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "kube-api-access-wpwnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.655678 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.662122 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts" (OuterVolumeSpecName: "scripts") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.675293 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data" (OuterVolumeSpecName: "config-data") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.675478 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.687133 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "85db256a-98d5-4bca-82ac-098d49b243e8" (UID: "85db256a-98d5-4bca-82ac-098d49b243e8"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.721541 4688 scope.go:117] "RemoveContainer" containerID="908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.733941 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.734354 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9c8p\" (UniqueName: \"kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735587 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735670 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735791 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735808 4688 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735819 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85db256a-98d5-4bca-82ac-098d49b243e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735828 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735840 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwnr\" (UniqueName: \"kubernetes.io/projected/85db256a-98d5-4bca-82ac-098d49b243e8-kube-api-access-wpwnr\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.735849 4688 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/85db256a-98d5-4bca-82ac-098d49b243e8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.750615 4688 scope.go:117] "RemoveContainer" containerID="de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.751173 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4\": container with ID starting with de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4 not found: ID does not exist" containerID="de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.751280 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4"} err="failed to get container status \"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4\": rpc error: code = NotFound desc = could not find container \"de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4\": container with ID starting with de90400ff72eb4c42192f456f2cfe891c3a555673fffcc52c6a3b688880494a4 not found: ID does not exist" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.751360 4688 scope.go:117] "RemoveContainer" containerID="908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.751900 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc\": container with ID starting with 908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc not found: ID does not exist" containerID="908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.751992 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc"} err="failed to get container status \"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc\": rpc error: code = NotFound desc = could not find container \"908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc\": container with ID starting with 908a11210447d5d2a2ddf750c44863e606900164b5a0ed08c40c48127a08c4dc not found: ID does not exist" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.752062 4688 scope.go:117] "RemoveContainer" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.806743 4688 scope.go:117] "RemoveContainer" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" Oct 01 16:07:03 crc kubenswrapper[4688]: E1001 16:07:03.807238 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560\": container with ID starting with 99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560 not found: ID does not exist" containerID="99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.807266 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560"} err="failed to get container status \"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560\": rpc error: code = NotFound desc = could not find container \"99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560\": container with ID starting with 99068836d38dcd22ed040420e609589b861ee57efde27f1a58dc7803b0510560 not found: ID does not exist" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.840298 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle\") pod \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.840810 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs\") pod \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.840936 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data\") pod \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841082 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plp9b\" (UniqueName: \"kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b\") pod \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\" (UID: \"d7e75ac4-87bc-431a-9ec5-943ec836a9bc\") " Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841444 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9c8p\" (UniqueName: \"kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841587 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs" (OuterVolumeSpecName: "logs") pod "d7e75ac4-87bc-431a-9ec5-943ec836a9bc" (UID: "d7e75ac4-87bc-431a-9ec5-943ec836a9bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841715 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841850 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.841962 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.854830 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b" (OuterVolumeSpecName: "kube-api-access-plp9b") pod "d7e75ac4-87bc-431a-9ec5-943ec836a9bc" (UID: "d7e75ac4-87bc-431a-9ec5-943ec836a9bc"). InnerVolumeSpecName "kube-api-access-plp9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.855440 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.856455 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.857014 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.858543 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9c8p\" (UniqueName: \"kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p\") pod \"nova-scheduler-0\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.864224 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-654b94b7dd-g2k9q"] Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.880073 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7e75ac4-87bc-431a-9ec5-943ec836a9bc" (UID: "d7e75ac4-87bc-431a-9ec5-943ec836a9bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.888601 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data" (OuterVolumeSpecName: "config-data") pod "d7e75ac4-87bc-431a-9ec5-943ec836a9bc" (UID: "d7e75ac4-87bc-431a-9ec5-943ec836a9bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.900008 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.944123 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plp9b\" (UniqueName: \"kubernetes.io/projected/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-kube-api-access-plp9b\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.944370 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:03 crc kubenswrapper[4688]: I1001 16:07:03.944380 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e75ac4-87bc-431a-9ec5-943ec836a9bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.376131 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.530659 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7e75ac4-87bc-431a-9ec5-943ec836a9bc","Type":"ContainerDied","Data":"0ce4f3165613fb17645072ad70d7d66415d5c155fe33255b9a1d3dafb0999257"} Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.531917 4688 scope.go:117] "RemoveContainer" containerID="0b89eaafca7a994efabf5e1551a31ac25d4fe24c8a2a6dc13ed90b43ec384afd" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.531620 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.535598 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264","Type":"ContainerStarted","Data":"fd431521c1c8e2d27a84910266d9921b8da645c5cb34dd4f929820bba5791864"} Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.561338 4688 scope.go:117] "RemoveContainer" containerID="3cf9a7790f2460c71b210d0810788a3901e37b6d665e94a08fc15561b399539c" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.597311 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.607831 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.615949 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:04 crc kubenswrapper[4688]: E1001 16:07:04.616404 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-api" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.616424 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-api" Oct 01 16:07:04 crc kubenswrapper[4688]: E1001 16:07:04.616454 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-log" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.616464 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-log" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.616721 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" containerName="horizon" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.616752 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-api" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.616767 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" containerName="nova-api-log" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.617987 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.622420 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.623072 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.656886 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.656951 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.657004 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.657047 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdsv\" (UniqueName: \"kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.758891 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdsv\" (UniqueName: \"kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.759195 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.759316 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.759446 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.760094 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.767027 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.767080 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.780147 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdsv\" (UniqueName: \"kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv\") pod \"nova-api-0\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " pod="openstack/nova-api-0" Oct 01 16:07:04 crc kubenswrapper[4688]: I1001 16:07:04.947387 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.432842 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469801bc-76a8-4241-a4e1-7fbd9bd9d39f" path="/var/lib/kubelet/pods/469801bc-76a8-4241-a4e1-7fbd9bd9d39f/volumes" Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.434121 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85db256a-98d5-4bca-82ac-098d49b243e8" path="/var/lib/kubelet/pods/85db256a-98d5-4bca-82ac-098d49b243e8/volumes" Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.435104 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e75ac4-87bc-431a-9ec5-943ec836a9bc" path="/var/lib/kubelet/pods/d7e75ac4-87bc-431a-9ec5-943ec836a9bc/volumes" Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.494378 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.561819 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerStarted","Data":"d9c1d171ddbe3dcb3f8e8769f1af84f101b8dbac123d568223153553976d7dc7"} Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.563856 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264","Type":"ContainerStarted","Data":"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf"} Oct 01 16:07:05 crc kubenswrapper[4688]: I1001 16:07:05.590260 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.590241206 podStartE2EDuration="2.590241206s" podCreationTimestamp="2025-10-01 16:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:05.583824064 +0000 UTC m=+1214.934464036" watchObservedRunningTime="2025-10-01 16:07:05.590241206 +0000 UTC m=+1214.940881178" Oct 01 16:07:06 crc kubenswrapper[4688]: I1001 16:07:06.586594 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerStarted","Data":"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a"} Oct 01 16:07:06 crc kubenswrapper[4688]: I1001 16:07:06.587740 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerStarted","Data":"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe"} Oct 01 16:07:06 crc kubenswrapper[4688]: I1001 16:07:06.628420 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.628401713 podStartE2EDuration="2.628401713s" podCreationTimestamp="2025-10-01 16:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:06.620267662 +0000 UTC m=+1215.970907674" watchObservedRunningTime="2025-10-01 16:07:06.628401713 +0000 UTC m=+1215.979041675" Oct 01 16:07:08 crc kubenswrapper[4688]: I1001 16:07:08.901015 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:07:09 crc kubenswrapper[4688]: I1001 16:07:09.842570 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 16:07:13 crc kubenswrapper[4688]: I1001 16:07:13.900985 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:07:13 crc kubenswrapper[4688]: I1001 16:07:13.932753 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:07:14 crc kubenswrapper[4688]: I1001 16:07:14.695919 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:07:14 crc kubenswrapper[4688]: I1001 16:07:14.948339 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:07:14 crc kubenswrapper[4688]: I1001 16:07:14.948398 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:07:16 crc kubenswrapper[4688]: I1001 16:07:16.030690 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:16 crc kubenswrapper[4688]: I1001 16:07:16.030929 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.726779 4688 generic.go:334] "Generic (PLEG): container finished" podID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerID="58f1884793ca8c394b4449d587709fe97e059bdbb69059104602956929edf013" exitCode=137 Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.726846 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerDied","Data":"58f1884793ca8c394b4449d587709fe97e059bdbb69059104602956929edf013"} Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.727136 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0","Type":"ContainerDied","Data":"10a3e08635ae5784c03e3e764a5865f3de75608ef8d3bf24dae6097e3320e2e0"} Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.727152 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10a3e08635ae5784c03e3e764a5865f3de75608ef8d3bf24dae6097e3320e2e0" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.729403 4688 generic.go:334] "Generic (PLEG): container finished" podID="1682b567-2a60-4625-a0cd-d44627b2f7ec" containerID="f87044b84436f2a688c3f949cb91e14f043f3c4c85739aade57d998dac4b2ff4" exitCode=137 Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.729431 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1682b567-2a60-4625-a0cd-d44627b2f7ec","Type":"ContainerDied","Data":"f87044b84436f2a688c3f949cb91e14f043f3c4c85739aade57d998dac4b2ff4"} Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.729446 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1682b567-2a60-4625-a0cd-d44627b2f7ec","Type":"ContainerDied","Data":"e05635b5fb5965e985007b6b052796d1e21468ccb090d5a21c69bfa6a144b61f"} Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.729456 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e05635b5fb5965e985007b6b052796d1e21468ccb090d5a21c69bfa6a144b61f" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.730929 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.738086 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.896680 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data\") pod \"1682b567-2a60-4625-a0cd-d44627b2f7ec\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.896765 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjzb7\" (UniqueName: \"kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7\") pod \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.896884 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs\") pod \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.896916 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle\") pod \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.896965 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data\") pod \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\" (UID: \"a8240169-b6cd-443e-b4c9-2b9f1b6a2df0\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.897006 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle\") pod \"1682b567-2a60-4625-a0cd-d44627b2f7ec\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.897034 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svrvd\" (UniqueName: \"kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd\") pod \"1682b567-2a60-4625-a0cd-d44627b2f7ec\" (UID: \"1682b567-2a60-4625-a0cd-d44627b2f7ec\") " Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.897457 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs" (OuterVolumeSpecName: "logs") pod "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" (UID: "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.903476 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd" (OuterVolumeSpecName: "kube-api-access-svrvd") pod "1682b567-2a60-4625-a0cd-d44627b2f7ec" (UID: "1682b567-2a60-4625-a0cd-d44627b2f7ec"). InnerVolumeSpecName "kube-api-access-svrvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.904694 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7" (OuterVolumeSpecName: "kube-api-access-hjzb7") pod "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" (UID: "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0"). InnerVolumeSpecName "kube-api-access-hjzb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.927856 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data" (OuterVolumeSpecName: "config-data") pod "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" (UID: "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.931423 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data" (OuterVolumeSpecName: "config-data") pod "1682b567-2a60-4625-a0cd-d44627b2f7ec" (UID: "1682b567-2a60-4625-a0cd-d44627b2f7ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.946896 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" (UID: "a8240169-b6cd-443e-b4c9-2b9f1b6a2df0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:21 crc kubenswrapper[4688]: I1001 16:07:21.959209 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1682b567-2a60-4625-a0cd-d44627b2f7ec" (UID: "1682b567-2a60-4625-a0cd-d44627b2f7ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000049 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000243 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjzb7\" (UniqueName: \"kubernetes.io/projected/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-kube-api-access-hjzb7\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000264 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000277 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000288 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000312 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1682b567-2a60-4625-a0cd-d44627b2f7ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.000380 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svrvd\" (UniqueName: \"kubernetes.io/projected/1682b567-2a60-4625-a0cd-d44627b2f7ec-kube-api-access-svrvd\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.739145 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.739221 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.783044 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.799294 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.813033 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.822806 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831310 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: E1001 16:07:22.831682 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-log" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831696 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-log" Oct 01 16:07:22 crc kubenswrapper[4688]: E1001 16:07:22.831710 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-metadata" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831718 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-metadata" Oct 01 16:07:22 crc kubenswrapper[4688]: E1001 16:07:22.831740 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1682b567-2a60-4625-a0cd-d44627b2f7ec" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831746 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="1682b567-2a60-4625-a0cd-d44627b2f7ec" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831916 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-metadata" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831929 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="1682b567-2a60-4625-a0cd-d44627b2f7ec" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.831946 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" containerName="nova-metadata-log" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.832555 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.835954 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.836183 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.836405 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.840223 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.842057 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.845274 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.845585 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.850942 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:07:22 crc kubenswrapper[4688]: I1001 16:07:22.860398 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.020660 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021018 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021120 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021154 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021180 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68lj9\" (UniqueName: \"kubernetes.io/projected/050f05f1-0f70-46e1-94a3-e32d2aa42c36-kube-api-access-68lj9\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021211 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021248 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021267 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwktn\" (UniqueName: \"kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021292 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.021310 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122764 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122814 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwktn\" (UniqueName: \"kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122842 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122860 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122899 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122935 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.122994 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.123019 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.123046 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68lj9\" (UniqueName: \"kubernetes.io/projected/050f05f1-0f70-46e1-94a3-e32d2aa42c36-kube-api-access-68lj9\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.123069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.124758 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.130695 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.131890 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.135428 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.135752 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.137980 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.140795 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68lj9\" (UniqueName: \"kubernetes.io/projected/050f05f1-0f70-46e1-94a3-e32d2aa42c36-kube-api-access-68lj9\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.141696 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.139892 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050f05f1-0f70-46e1-94a3-e32d2aa42c36-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"050f05f1-0f70-46e1-94a3-e32d2aa42c36\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.159818 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.161943 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwktn\" (UniqueName: \"kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn\") pod \"nova-metadata-0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.177978 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.391346 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1682b567-2a60-4625-a0cd-d44627b2f7ec" path="/var/lib/kubelet/pods/1682b567-2a60-4625-a0cd-d44627b2f7ec/volumes" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.392157 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8240169-b6cd-443e-b4c9-2b9f1b6a2df0" path="/var/lib/kubelet/pods/a8240169-b6cd-443e-b4c9-2b9f1b6a2df0/volumes" Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.663654 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.724880 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 16:07:23 crc kubenswrapper[4688]: W1001 16:07:23.733918 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod050f05f1_0f70_46e1_94a3_e32d2aa42c36.slice/crio-53b5746f275b84e59461ba0b74d7308722fe495bf6942031b5875dd320011107 WatchSource:0}: Error finding container 53b5746f275b84e59461ba0b74d7308722fe495bf6942031b5875dd320011107: Status 404 returned error can't find the container with id 53b5746f275b84e59461ba0b74d7308722fe495bf6942031b5875dd320011107 Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.749321 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"050f05f1-0f70-46e1-94a3-e32d2aa42c36","Type":"ContainerStarted","Data":"53b5746f275b84e59461ba0b74d7308722fe495bf6942031b5875dd320011107"} Oct 01 16:07:23 crc kubenswrapper[4688]: I1001 16:07:23.753629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerStarted","Data":"eedfbab440da2cbc043d610a50b5a2f1063933db18f186e6053f2e20fcda145d"} Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.762298 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"050f05f1-0f70-46e1-94a3-e32d2aa42c36","Type":"ContainerStarted","Data":"9af867396261a92e37388d23978ce88d8b78cdfad140a6a2a0fa4d9a78b70029"} Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.766873 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerStarted","Data":"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4"} Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.767822 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerStarted","Data":"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e"} Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.784641 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.784620601 podStartE2EDuration="2.784620601s" podCreationTimestamp="2025-10-01 16:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:24.780031751 +0000 UTC m=+1234.130671723" watchObservedRunningTime="2025-10-01 16:07:24.784620601 +0000 UTC m=+1234.135260563" Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.806831 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.806815803 podStartE2EDuration="2.806815803s" podCreationTimestamp="2025-10-01 16:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:24.804511067 +0000 UTC m=+1234.155151029" watchObservedRunningTime="2025-10-01 16:07:24.806815803 +0000 UTC m=+1234.157455765" Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.952756 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.953330 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.956000 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:07:24 crc kubenswrapper[4688]: I1001 16:07:24.960477 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:07:25 crc kubenswrapper[4688]: I1001 16:07:25.777124 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:07:25 crc kubenswrapper[4688]: I1001 16:07:25.780979 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:07:25 crc kubenswrapper[4688]: I1001 16:07:25.981872 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:07:25 crc kubenswrapper[4688]: I1001 16:07:25.998978 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:25 crc kubenswrapper[4688]: I1001 16:07:25.999448 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101531 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101588 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101630 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101661 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101782 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.101801 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsc5d\" (UniqueName: \"kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.204848 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsc5d\" (UniqueName: \"kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.204910 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.204932 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.204955 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.204987 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.205116 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.207207 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.207861 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.207937 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.208127 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.208292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.257376 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsc5d\" (UniqueName: \"kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d\") pod \"dnsmasq-dns-cd5cbd7b9-8vh5d\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.328212 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:26 crc kubenswrapper[4688]: I1001 16:07:26.836584 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:07:27 crc kubenswrapper[4688]: I1001 16:07:27.392186 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 16:07:27 crc kubenswrapper[4688]: I1001 16:07:27.795880 4688 generic.go:334] "Generic (PLEG): container finished" podID="ea25de60-738d-40ce-bc03-37155b616e1e" containerID="634215ffb6afedf22db1c31fe24c482376cbd5dd700f4f6ee15d5d9c836ac1e2" exitCode=0 Oct 01 16:07:27 crc kubenswrapper[4688]: I1001 16:07:27.795929 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" event={"ID":"ea25de60-738d-40ce-bc03-37155b616e1e","Type":"ContainerDied","Data":"634215ffb6afedf22db1c31fe24c482376cbd5dd700f4f6ee15d5d9c836ac1e2"} Oct 01 16:07:27 crc kubenswrapper[4688]: I1001 16:07:27.795978 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" event={"ID":"ea25de60-738d-40ce-bc03-37155b616e1e","Type":"ContainerStarted","Data":"dc25c7734b62312080358dc57a63bb65ee919ed7cfaead1e2641619c93f5bd2d"} Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.161075 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.178764 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.178809 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.277677 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.277917 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-central-agent" containerID="cri-o://47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.278025 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="proxy-httpd" containerID="cri-o://8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.278084 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-notification-agent" containerID="cri-o://53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.278121 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="sg-core" containerID="cri-o://3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.518932 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806498 4688 generic.go:334] "Generic (PLEG): container finished" podID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerID="8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b" exitCode=0 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806752 4688 generic.go:334] "Generic (PLEG): container finished" podID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerID="3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1" exitCode=2 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806760 4688 generic.go:334] "Generic (PLEG): container finished" podID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerID="47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd" exitCode=0 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806561 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerDied","Data":"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b"} Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806826 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerDied","Data":"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1"} Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.806857 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerDied","Data":"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd"} Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.808807 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" event={"ID":"ea25de60-738d-40ce-bc03-37155b616e1e","Type":"ContainerStarted","Data":"f4a78be3c595360ad8f3085e62a0ea1c42e1ad7f88737f5a98b22f4d4a7b7fc3"} Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.808932 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.809125 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-log" containerID="cri-o://aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.809630 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-api" containerID="cri-o://057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a" gracePeriod=30 Oct 01 16:07:28 crc kubenswrapper[4688]: I1001 16:07:28.848606 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" podStartSLOduration=3.848586765 podStartE2EDuration="3.848586765s" podCreationTimestamp="2025-10-01 16:07:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:28.839008743 +0000 UTC m=+1238.189648705" watchObservedRunningTime="2025-10-01 16:07:28.848586765 +0000 UTC m=+1238.199226727" Oct 01 16:07:29 crc kubenswrapper[4688]: I1001 16:07:29.823221 4688 generic.go:334] "Generic (PLEG): container finished" podID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerID="aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe" exitCode=143 Oct 01 16:07:29 crc kubenswrapper[4688]: I1001 16:07:29.823669 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerDied","Data":"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe"} Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.799745 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.842369 4688 generic.go:334] "Generic (PLEG): container finished" podID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerID="53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32" exitCode=0 Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.842421 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerDied","Data":"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32"} Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.842905 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a9f48e4-a0e0-4227-b108-8caea9529999","Type":"ContainerDied","Data":"c9cca9d35829cf7ea995f3c1e50ac891396d697441aaddbda3f1b4867037e90c"} Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.842467 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.842975 4688 scope.go:117] "RemoveContainer" containerID="8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.885881 4688 scope.go:117] "RemoveContainer" containerID="3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.908636 4688 scope.go:117] "RemoveContainer" containerID="53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.936894 4688 scope.go:117] "RemoveContainer" containerID="47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.953365 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.953405 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.953527 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.954113 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.954142 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.954167 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.954375 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.954416 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mngg\" (UniqueName: \"kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg\") pod \"0a9f48e4-a0e0-4227-b108-8caea9529999\" (UID: \"0a9f48e4-a0e0-4227-b108-8caea9529999\") " Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.955238 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.955977 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.962033 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg" (OuterVolumeSpecName: "kube-api-access-6mngg") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "kube-api-access-6mngg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.964060 4688 scope.go:117] "RemoveContainer" containerID="8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b" Oct 01 16:07:31 crc kubenswrapper[4688]: E1001 16:07:31.964396 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b\": container with ID starting with 8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b not found: ID does not exist" containerID="8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.964421 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b"} err="failed to get container status \"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b\": rpc error: code = NotFound desc = could not find container \"8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b\": container with ID starting with 8d5ab71fbbf7ed685c5577152b4a9cd9117b69b4841380ca69ef44758f9fd58b not found: ID does not exist" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.964441 4688 scope.go:117] "RemoveContainer" containerID="3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1" Oct 01 16:07:31 crc kubenswrapper[4688]: E1001 16:07:31.964878 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1\": container with ID starting with 3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1 not found: ID does not exist" containerID="3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.964899 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1"} err="failed to get container status \"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1\": rpc error: code = NotFound desc = could not find container \"3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1\": container with ID starting with 3110672a484f9bcc7f9212c36a1eee62d3a4fa0f7b5f16409e465ed7ca08ead1 not found: ID does not exist" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.964912 4688 scope.go:117] "RemoveContainer" containerID="53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32" Oct 01 16:07:31 crc kubenswrapper[4688]: E1001 16:07:31.965148 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32\": container with ID starting with 53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32 not found: ID does not exist" containerID="53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.965165 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32"} err="failed to get container status \"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32\": rpc error: code = NotFound desc = could not find container \"53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32\": container with ID starting with 53da3f34df724000604a4c5a0f20eac6b83875ed6f49eea9425ab347ab0f3f32 not found: ID does not exist" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.965176 4688 scope.go:117] "RemoveContainer" containerID="47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd" Oct 01 16:07:31 crc kubenswrapper[4688]: E1001 16:07:31.965571 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd\": container with ID starting with 47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd not found: ID does not exist" containerID="47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.965588 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd"} err="failed to get container status \"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd\": rpc error: code = NotFound desc = could not find container \"47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd\": container with ID starting with 47ac66ced0a6d0fb97e087eef3928d9508699cd3de18fb22401cb556c3d1e8bd not found: ID does not exist" Oct 01 16:07:31 crc kubenswrapper[4688]: I1001 16:07:31.975969 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts" (OuterVolumeSpecName: "scripts") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.058263 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a9f48e4-a0e0-4227-b108-8caea9529999-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.058295 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mngg\" (UniqueName: \"kubernetes.io/projected/0a9f48e4-a0e0-4227-b108-8caea9529999-kube-api-access-6mngg\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.058306 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.067988 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.092707 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data" (OuterVolumeSpecName: "config-data") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.107987 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a9f48e4-a0e0-4227-b108-8caea9529999" (UID: "0a9f48e4-a0e0-4227-b108-8caea9529999"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.160505 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.161472 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.161484 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9f48e4-a0e0-4227-b108-8caea9529999-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.215579 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.242672 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.253850 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:32 crc kubenswrapper[4688]: E1001 16:07:32.254284 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="proxy-httpd" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254306 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="proxy-httpd" Oct 01 16:07:32 crc kubenswrapper[4688]: E1001 16:07:32.254319 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-notification-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254325 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-notification-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: E1001 16:07:32.254345 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="sg-core" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254354 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="sg-core" Oct 01 16:07:32 crc kubenswrapper[4688]: E1001 16:07:32.254364 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-central-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254370 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-central-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254605 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-notification-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254620 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="sg-core" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254651 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="ceilometer-central-agent" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.254676 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" containerName="proxy-httpd" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.258487 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.263478 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.263710 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.264984 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265060 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265102 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265146 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265174 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6qvn\" (UniqueName: \"kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265254 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.265279 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.266062 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366104 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366183 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366208 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366257 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6qvn\" (UniqueName: \"kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366277 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.366351 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.367586 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.367439 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.367324 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.371425 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.371465 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.372560 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.374424 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.391756 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6qvn\" (UniqueName: \"kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn\") pod \"ceilometer-0\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.488840 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.571923 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle\") pod \"3a2439d1-cbf2-424e-820a-738ebe80f59a\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.572037 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcdsv\" (UniqueName: \"kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv\") pod \"3a2439d1-cbf2-424e-820a-738ebe80f59a\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.572090 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs\") pod \"3a2439d1-cbf2-424e-820a-738ebe80f59a\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.572129 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data\") pod \"3a2439d1-cbf2-424e-820a-738ebe80f59a\" (UID: \"3a2439d1-cbf2-424e-820a-738ebe80f59a\") " Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.572857 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs" (OuterVolumeSpecName: "logs") pod "3a2439d1-cbf2-424e-820a-738ebe80f59a" (UID: "3a2439d1-cbf2-424e-820a-738ebe80f59a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.611914 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv" (OuterVolumeSpecName: "kube-api-access-hcdsv") pod "3a2439d1-cbf2-424e-820a-738ebe80f59a" (UID: "3a2439d1-cbf2-424e-820a-738ebe80f59a"). InnerVolumeSpecName "kube-api-access-hcdsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.646963 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data" (OuterVolumeSpecName: "config-data") pod "3a2439d1-cbf2-424e-820a-738ebe80f59a" (UID: "3a2439d1-cbf2-424e-820a-738ebe80f59a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.652369 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.673849 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2439d1-cbf2-424e-820a-738ebe80f59a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.673870 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.673880 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcdsv\" (UniqueName: \"kubernetes.io/projected/3a2439d1-cbf2-424e-820a-738ebe80f59a-kube-api-access-hcdsv\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.676389 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a2439d1-cbf2-424e-820a-738ebe80f59a" (UID: "3a2439d1-cbf2-424e-820a-738ebe80f59a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.776882 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2439d1-cbf2-424e-820a-738ebe80f59a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.891361 4688 generic.go:334] "Generic (PLEG): container finished" podID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerID="057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a" exitCode=0 Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.891412 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerDied","Data":"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a"} Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.891441 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a2439d1-cbf2-424e-820a-738ebe80f59a","Type":"ContainerDied","Data":"d9c1d171ddbe3dcb3f8e8769f1af84f101b8dbac123d568223153553976d7dc7"} Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.891457 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.891462 4688 scope.go:117] "RemoveContainer" containerID="057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.930879 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.973635 4688 scope.go:117] "RemoveContainer" containerID="aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe" Oct 01 16:07:32 crc kubenswrapper[4688]: I1001 16:07:32.998018 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.020756 4688 scope.go:117] "RemoveContainer" containerID="057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a" Oct 01 16:07:33 crc kubenswrapper[4688]: E1001 16:07:33.026609 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a\": container with ID starting with 057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a not found: ID does not exist" containerID="057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.026657 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a"} err="failed to get container status \"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a\": rpc error: code = NotFound desc = could not find container \"057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a\": container with ID starting with 057241cac80f7c1641642ab36402a0ba9c7ccf857150755e5ca6d5c90c67029a not found: ID does not exist" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.026685 4688 scope.go:117] "RemoveContainer" containerID="aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe" Oct 01 16:07:33 crc kubenswrapper[4688]: E1001 16:07:33.029643 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe\": container with ID starting with aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe not found: ID does not exist" containerID="aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.029673 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe"} err="failed to get container status \"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe\": rpc error: code = NotFound desc = could not find container \"aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe\": container with ID starting with aaf533d0d9c81517cf4be86d2c5062a5c6251215b128138a4c11c3477c741ffe not found: ID does not exist" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.029707 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: E1001 16:07:33.030401 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-log" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.030413 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-log" Oct 01 16:07:33 crc kubenswrapper[4688]: E1001 16:07:33.030454 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-api" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.030460 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-api" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.030890 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-api" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.030913 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" containerName="nova-api-log" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.032275 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.037940 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.038065 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.041434 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.046494 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.161750 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.179057 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.179102 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.188866 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.188959 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.189091 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.189150 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4hcp\" (UniqueName: \"kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.189185 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.189323 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.191145 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.277834 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.282325 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291374 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291436 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4hcp\" (UniqueName: \"kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291459 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291579 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291668 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.291697 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.293088 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.301050 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.302095 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.303419 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.308241 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.322329 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4hcp\" (UniqueName: \"kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp\") pod \"nova-api-0\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.370609 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.415228 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9f48e4-a0e0-4227-b108-8caea9529999" path="/var/lib/kubelet/pods/0a9f48e4-a0e0-4227-b108-8caea9529999/volumes" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.417304 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2439d1-cbf2-424e-820a-738ebe80f59a" path="/var/lib/kubelet/pods/3a2439d1-cbf2-424e-820a-738ebe80f59a/volumes" Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.551626 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.552100 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" containerName="kube-state-metrics" containerID="cri-o://b2b6b7f7ae0da18ce6e06e3f7e7c4c90ad06ef8c409d8ea5b3e27d43597ee551" gracePeriod=30 Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.936861 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerStarted","Data":"47626915c8a75115c02657ba490873a468a0e09400c159b6b7670ae58fe495d6"} Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.938596 4688 generic.go:334] "Generic (PLEG): container finished" podID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" containerID="b2b6b7f7ae0da18ce6e06e3f7e7c4c90ad06ef8c409d8ea5b3e27d43597ee551" exitCode=2 Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.942589 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5","Type":"ContainerDied","Data":"b2b6b7f7ae0da18ce6e06e3f7e7c4c90ad06ef8c409d8ea5b3e27d43597ee551"} Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.944306 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:33 crc kubenswrapper[4688]: I1001 16:07:33.986728 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.111792 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.191688 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.191945 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.207089 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjx74\" (UniqueName: \"kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74\") pod \"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5\" (UID: \"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5\") " Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.215315 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74" (OuterVolumeSpecName: "kube-api-access-zjx74") pod "9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" (UID: "9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5"). InnerVolumeSpecName "kube-api-access-zjx74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.253830 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rbx2w"] Oct 01 16:07:34 crc kubenswrapper[4688]: E1001 16:07:34.254432 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" containerName="kube-state-metrics" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.254505 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" containerName="kube-state-metrics" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.254790 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" containerName="kube-state-metrics" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.255438 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.258774 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.259412 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.271148 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbx2w"] Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.309884 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjx74\" (UniqueName: \"kubernetes.io/projected/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5-kube-api-access-zjx74\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.412525 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.416803 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdjk8\" (UniqueName: \"kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.417038 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.417109 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.518240 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.518313 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.518334 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.518355 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdjk8\" (UniqueName: \"kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.531232 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.543327 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.548590 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdjk8\" (UniqueName: \"kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.556069 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts\") pod \"nova-cell1-cell-mapping-rbx2w\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.593930 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.950111 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerStarted","Data":"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980"} Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.975019 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerStarted","Data":"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830"} Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.975061 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerStarted","Data":"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537"} Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.975069 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerStarted","Data":"2de1718b4bc75300a9b80c2a6e6d33459ea88380a6c87f23307086eb8feb0a20"} Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.979349 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.979407 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5","Type":"ContainerDied","Data":"97118b400b502785273900cb1191594ce1da9a9d1869d8b80705aecbcc4a12f1"} Oct 01 16:07:34 crc kubenswrapper[4688]: I1001 16:07:34.979443 4688 scope.go:117] "RemoveContainer" containerID="b2b6b7f7ae0da18ce6e06e3f7e7c4c90ad06ef8c409d8ea5b3e27d43597ee551" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.023795 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.023774905 podStartE2EDuration="3.023774905s" podCreationTimestamp="2025-10-01 16:07:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:35.005185956 +0000 UTC m=+1244.355825918" watchObservedRunningTime="2025-10-01 16:07:35.023774905 +0000 UTC m=+1244.374414867" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.065929 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.075390 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.097373 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.098560 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.102932 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.103129 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.104402 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.180241 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbx2w"] Oct 01 16:07:35 crc kubenswrapper[4688]: W1001 16:07:35.180438 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8199e7d8_c362_43ae_a734_d287ff9f293d.slice/crio-f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad WatchSource:0}: Error finding container f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad: Status 404 returned error can't find the container with id f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.232470 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.232515 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.232562 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcs4w\" (UniqueName: \"kubernetes.io/projected/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-api-access-fcs4w\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.232605 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.334385 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.334566 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.334620 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.334654 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcs4w\" (UniqueName: \"kubernetes.io/projected/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-api-access-fcs4w\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.341429 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.341784 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.348147 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.356998 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcs4w\" (UniqueName: \"kubernetes.io/projected/c522fa53-fdbd-4836-b127-a858c5001cc0-kube-api-access-fcs4w\") pod \"kube-state-metrics-0\" (UID: \"c522fa53-fdbd-4836-b127-a858c5001cc0\") " pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.404294 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5" path="/var/lib/kubelet/pods/9ab594f4-2b07-4d58-bb14-b2ef7dbd59c5/volumes" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.430112 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 16:07:35 crc kubenswrapper[4688]: I1001 16:07:35.795615 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:35.992082 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c522fa53-fdbd-4836-b127-a858c5001cc0","Type":"ContainerStarted","Data":"b8d4b0adf33ee97be1d53ccf35d9b0861c8b2a3d6902bf5fc3510a05451bae37"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.005990 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerStarted","Data":"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.007567 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbx2w" event={"ID":"8199e7d8-c362-43ae-a734-d287ff9f293d","Type":"ContainerStarted","Data":"86e31a9a079694bb042db75db204c2f0d3ede0e67af2ff1083533099cfa214dd"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.007607 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbx2w" event={"ID":"8199e7d8-c362-43ae-a734-d287ff9f293d","Type":"ContainerStarted","Data":"f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.035607 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rbx2w" podStartSLOduration=2.035590411 podStartE2EDuration="2.035590411s" podCreationTimestamp="2025-10-01 16:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:36.029604061 +0000 UTC m=+1245.380244023" watchObservedRunningTime="2025-10-01 16:07:36.035590411 +0000 UTC m=+1245.386230373" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.329696 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.421661 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:36.429235 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="dnsmasq-dns" containerID="cri-o://94373c3c90ca14b126a9b2291e350ba725ec647d2250b0255943597ed46e07ca" gracePeriod=10 Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.027478 4688 generic.go:334] "Generic (PLEG): container finished" podID="51006f0d-a088-4424-be58-fb677966614d" containerID="94373c3c90ca14b126a9b2291e350ba725ec647d2250b0255943597ed46e07ca" exitCode=0 Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.027756 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" event={"ID":"51006f0d-a088-4424-be58-fb677966614d","Type":"ContainerDied","Data":"94373c3c90ca14b126a9b2291e350ba725ec647d2250b0255943597ed46e07ca"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.029668 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerStarted","Data":"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503"} Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.537061 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681278 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681341 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8jpj\" (UniqueName: \"kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681464 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681498 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681549 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.681578 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config\") pod \"51006f0d-a088-4424-be58-fb677966614d\" (UID: \"51006f0d-a088-4424-be58-fb677966614d\") " Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.697726 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj" (OuterVolumeSpecName: "kube-api-access-w8jpj") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "kube-api-access-w8jpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.778127 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.788936 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.788963 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8jpj\" (UniqueName: \"kubernetes.io/projected/51006f0d-a088-4424-be58-fb677966614d-kube-api-access-w8jpj\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.862024 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config" (OuterVolumeSpecName: "config") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.879055 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.879680 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.887932 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51006f0d-a088-4424-be58-fb677966614d" (UID: "51006f0d-a088-4424-be58-fb677966614d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.890444 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.890471 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.890483 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.890493 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51006f0d-a088-4424-be58-fb677966614d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:37 crc kubenswrapper[4688]: I1001 16:07:37.946814 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.040048 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" event={"ID":"51006f0d-a088-4424-be58-fb677966614d","Type":"ContainerDied","Data":"39d103a06934eaab47bd10275909b8966be4870e071bdc38bd642a99315018e7"} Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.040100 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-dj9ns" Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.040439 4688 scope.go:117] "RemoveContainer" containerID="94373c3c90ca14b126a9b2291e350ba725ec647d2250b0255943597ed46e07ca" Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.041981 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c522fa53-fdbd-4836-b127-a858c5001cc0","Type":"ContainerStarted","Data":"99a5bc98892b3aa095b51d50a6dea5ea11e64a0a9cfdecc2168ca1d1cd03235f"} Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.042904 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.066094 4688 scope.go:117] "RemoveContainer" containerID="07d2c36542a2d4d8109f9b3985d9399d72d8d66a2a2551cba27a4a6f4bb5d693" Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.066091 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9864268539999999 podStartE2EDuration="3.066073831s" podCreationTimestamp="2025-10-01 16:07:35 +0000 UTC" firstStartedPulling="2025-10-01 16:07:35.800200824 +0000 UTC m=+1245.150840796" lastFinishedPulling="2025-10-01 16:07:36.879847811 +0000 UTC m=+1246.230487773" observedRunningTime="2025-10-01 16:07:38.063318662 +0000 UTC m=+1247.413958624" watchObservedRunningTime="2025-10-01 16:07:38.066073831 +0000 UTC m=+1247.416713803" Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.092140 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:07:38 crc kubenswrapper[4688]: I1001 16:07:38.101405 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-dj9ns"] Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067092 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerStarted","Data":"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f"} Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067509 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067503 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="proxy-httpd" containerID="cri-o://41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f" gracePeriod=30 Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067573 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="sg-core" containerID="cri-o://3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503" gracePeriod=30 Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067644 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-notification-agent" containerID="cri-o://e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb" gracePeriod=30 Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.067897 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-central-agent" containerID="cri-o://a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980" gracePeriod=30 Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.101105 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.258396158 podStartE2EDuration="7.101085547s" podCreationTimestamp="2025-10-01 16:07:32 +0000 UTC" firstStartedPulling="2025-10-01 16:07:33.277649746 +0000 UTC m=+1242.628289708" lastFinishedPulling="2025-10-01 16:07:38.120339135 +0000 UTC m=+1247.470979097" observedRunningTime="2025-10-01 16:07:39.097592278 +0000 UTC m=+1248.448232240" watchObservedRunningTime="2025-10-01 16:07:39.101085547 +0000 UTC m=+1248.451725509" Oct 01 16:07:39 crc kubenswrapper[4688]: I1001 16:07:39.391384 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51006f0d-a088-4424-be58-fb677966614d" path="/var/lib/kubelet/pods/51006f0d-a088-4424-be58-fb677966614d/volumes" Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090403 4688 generic.go:334] "Generic (PLEG): container finished" podID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerID="41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f" exitCode=0 Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090753 4688 generic.go:334] "Generic (PLEG): container finished" podID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerID="3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503" exitCode=2 Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090766 4688 generic.go:334] "Generic (PLEG): container finished" podID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerID="e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb" exitCode=0 Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090447 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerDied","Data":"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f"} Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090807 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerDied","Data":"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503"} Oct 01 16:07:40 crc kubenswrapper[4688]: I1001 16:07:40.090827 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerDied","Data":"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb"} Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.109272 4688 generic.go:334] "Generic (PLEG): container finished" podID="8199e7d8-c362-43ae-a734-d287ff9f293d" containerID="86e31a9a079694bb042db75db204c2f0d3ede0e67af2ff1083533099cfa214dd" exitCode=0 Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.109347 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbx2w" event={"ID":"8199e7d8-c362-43ae-a734-d287ff9f293d","Type":"ContainerDied","Data":"86e31a9a079694bb042db75db204c2f0d3ede0e67af2ff1083533099cfa214dd"} Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.547032 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696289 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696372 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696434 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696455 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696486 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696556 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.696582 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6qvn\" (UniqueName: \"kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn\") pod \"dcc58017-6d75-42cb-bfdc-61793dfad551\" (UID: \"dcc58017-6d75-42cb-bfdc-61793dfad551\") " Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.697262 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.697564 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.703149 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts" (OuterVolumeSpecName: "scripts") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.711083 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn" (OuterVolumeSpecName: "kube-api-access-m6qvn") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "kube-api-access-m6qvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.746710 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.801670 4688 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.801853 4688 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.801967 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.802038 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6qvn\" (UniqueName: \"kubernetes.io/projected/dcc58017-6d75-42cb-bfdc-61793dfad551-kube-api-access-m6qvn\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.802157 4688 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dcc58017-6d75-42cb-bfdc-61793dfad551-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.805999 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.828558 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data" (OuterVolumeSpecName: "config-data") pod "dcc58017-6d75-42cb-bfdc-61793dfad551" (UID: "dcc58017-6d75-42cb-bfdc-61793dfad551"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.903894 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:42 crc kubenswrapper[4688]: I1001 16:07:42.903928 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc58017-6d75-42cb-bfdc-61793dfad551-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.125227 4688 generic.go:334] "Generic (PLEG): container finished" podID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerID="a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980" exitCode=0 Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.125626 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.125654 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerDied","Data":"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980"} Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.125695 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dcc58017-6d75-42cb-bfdc-61793dfad551","Type":"ContainerDied","Data":"47626915c8a75115c02657ba490873a468a0e09400c159b6b7670ae58fe495d6"} Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.125718 4688 scope.go:117] "RemoveContainer" containerID="41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.179599 4688 scope.go:117] "RemoveContainer" containerID="3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.187725 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.200020 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.203066 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.222485 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223084 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="proxy-httpd" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223170 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="proxy-httpd" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223227 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-notification-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223273 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-notification-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223326 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="dnsmasq-dns" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223369 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="dnsmasq-dns" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223425 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-central-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223476 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-central-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223548 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="init" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223595 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="init" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.223651 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="sg-core" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223694 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="sg-core" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223897 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-notification-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.223960 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="proxy-httpd" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.224019 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="51006f0d-a088-4424-be58-fb677966614d" containerName="dnsmasq-dns" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.224068 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="sg-core" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.224121 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" containerName="ceilometer-central-agent" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.226366 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.230857 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.231581 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.232611 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.232824 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.236352 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.248026 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.299403 4688 scope.go:117] "RemoveContainer" containerID="e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.308982 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-log-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309042 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309085 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309117 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309136 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8mf6\" (UniqueName: \"kubernetes.io/projected/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-kube-api-access-n8mf6\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309166 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-scripts\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309198 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-run-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.309227 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-config-data\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.318715 4688 scope.go:117] "RemoveContainer" containerID="a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.345412 4688 scope.go:117] "RemoveContainer" containerID="41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.346011 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f\": container with ID starting with 41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f not found: ID does not exist" containerID="41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.346046 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f"} err="failed to get container status \"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f\": rpc error: code = NotFound desc = could not find container \"41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f\": container with ID starting with 41ca51798a599a8b662898bf97e74d5a5cc71eb0496d92c054e67ec57d37b67f not found: ID does not exist" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.346085 4688 scope.go:117] "RemoveContainer" containerID="3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.347594 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503\": container with ID starting with 3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503 not found: ID does not exist" containerID="3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.347647 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503"} err="failed to get container status \"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503\": rpc error: code = NotFound desc = could not find container \"3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503\": container with ID starting with 3faec7af52db71f7201c33c85dca1173fd52eb452ac37bff2696f6f43b229503 not found: ID does not exist" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.347664 4688 scope.go:117] "RemoveContainer" containerID="e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.348815 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb\": container with ID starting with e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb not found: ID does not exist" containerID="e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.348853 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb"} err="failed to get container status \"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb\": rpc error: code = NotFound desc = could not find container \"e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb\": container with ID starting with e94143ba7f91a03122d67c11dd79cdf28388b90e4376e7eb9a5aea8443f932fb not found: ID does not exist" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.348880 4688 scope.go:117] "RemoveContainer" containerID="a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980" Oct 01 16:07:43 crc kubenswrapper[4688]: E1001 16:07:43.349197 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980\": container with ID starting with a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980 not found: ID does not exist" containerID="a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.349213 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980"} err="failed to get container status \"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980\": rpc error: code = NotFound desc = could not find container \"a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980\": container with ID starting with a38fd9b64f79a142f8017d24d6769cfc8c480f3c5c96f126fe2c2fa59954a980 not found: ID does not exist" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.370977 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.371396 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.407654 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcc58017-6d75-42cb-bfdc-61793dfad551" path="/var/lib/kubelet/pods/dcc58017-6d75-42cb-bfdc-61793dfad551/volumes" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.412662 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.412969 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413068 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413203 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8mf6\" (UniqueName: \"kubernetes.io/projected/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-kube-api-access-n8mf6\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413294 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-scripts\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413389 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-run-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413572 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-config-data\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.413745 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-log-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.428448 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-log-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.429117 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-run-httpd\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.433317 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-scripts\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.437246 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.439746 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8mf6\" (UniqueName: \"kubernetes.io/projected/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-kube-api-access-n8mf6\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.443825 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.457596 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.457864 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab269f92-11bd-4ea1-82a3-a77c15c5bb07-config-data\") pod \"ceilometer-0\" (UID: \"ab269f92-11bd-4ea1-82a3-a77c15c5bb07\") " pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.577865 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.688507 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.821319 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdjk8\" (UniqueName: \"kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8\") pod \"8199e7d8-c362-43ae-a734-d287ff9f293d\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.821610 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts\") pod \"8199e7d8-c362-43ae-a734-d287ff9f293d\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.821648 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle\") pod \"8199e7d8-c362-43ae-a734-d287ff9f293d\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.821702 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data\") pod \"8199e7d8-c362-43ae-a734-d287ff9f293d\" (UID: \"8199e7d8-c362-43ae-a734-d287ff9f293d\") " Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.826881 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts" (OuterVolumeSpecName: "scripts") pod "8199e7d8-c362-43ae-a734-d287ff9f293d" (UID: "8199e7d8-c362-43ae-a734-d287ff9f293d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.831768 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8" (OuterVolumeSpecName: "kube-api-access-vdjk8") pod "8199e7d8-c362-43ae-a734-d287ff9f293d" (UID: "8199e7d8-c362-43ae-a734-d287ff9f293d"). InnerVolumeSpecName "kube-api-access-vdjk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.850574 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data" (OuterVolumeSpecName: "config-data") pod "8199e7d8-c362-43ae-a734-d287ff9f293d" (UID: "8199e7d8-c362-43ae-a734-d287ff9f293d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.852288 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8199e7d8-c362-43ae-a734-d287ff9f293d" (UID: "8199e7d8-c362-43ae-a734-d287ff9f293d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.924751 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.924791 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdjk8\" (UniqueName: \"kubernetes.io/projected/8199e7d8-c362-43ae-a734-d287ff9f293d-kube-api-access-vdjk8\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.924806 4688 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:43 crc kubenswrapper[4688]: I1001 16:07:43.924818 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8199e7d8-c362-43ae-a734-d287ff9f293d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:44 crc kubenswrapper[4688]: W1001 16:07:44.039377 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab269f92_11bd_4ea1_82a3_a77c15c5bb07.slice/crio-641324a9cf5e895ec914daf6f342ef0588ec188ff328ab42772acedd6d58682c WatchSource:0}: Error finding container 641324a9cf5e895ec914daf6f342ef0588ec188ff328ab42772acedd6d58682c: Status 404 returned error can't find the container with id 641324a9cf5e895ec914daf6f342ef0588ec188ff328ab42772acedd6d58682c Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.048285 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.137866 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab269f92-11bd-4ea1-82a3-a77c15c5bb07","Type":"ContainerStarted","Data":"641324a9cf5e895ec914daf6f342ef0588ec188ff328ab42772acedd6d58682c"} Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.146247 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rbx2w" Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.147803 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rbx2w" event={"ID":"8199e7d8-c362-43ae-a734-d287ff9f293d","Type":"ContainerDied","Data":"f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad"} Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.147841 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0462efdb7325c1498ee8cf6bc38ab827fdb7a5d76befed7d8bf6755f534efad" Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.164506 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.317610 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.366867 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.367061 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" containerName="nova-scheduler-scheduler" containerID="cri-o://5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf" gracePeriod=30 Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.379518 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.387666 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:07:44 crc kubenswrapper[4688]: I1001 16:07:44.390179 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:45 crc kubenswrapper[4688]: I1001 16:07:45.156478 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab269f92-11bd-4ea1-82a3-a77c15c5bb07","Type":"ContainerStarted","Data":"8c85757673a728d1af28b557b7f1d1f1fddbbec122a91c8fd8da2ad4786cca06"} Oct 01 16:07:45 crc kubenswrapper[4688]: I1001 16:07:45.156617 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-log" containerID="cri-o://fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537" gracePeriod=30 Oct 01 16:07:45 crc kubenswrapper[4688]: I1001 16:07:45.156686 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-api" containerID="cri-o://79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830" gracePeriod=30 Oct 01 16:07:45 crc kubenswrapper[4688]: I1001 16:07:45.457316 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.166980 4688 generic.go:334] "Generic (PLEG): container finished" podID="a062e1e2-1bd3-4543-8513-991de8af5608" containerID="fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537" exitCode=143 Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.167059 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerDied","Data":"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537"} Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.169962 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab269f92-11bd-4ea1-82a3-a77c15c5bb07","Type":"ContainerStarted","Data":"2680ed09d5aab221e59b86621f7e2fd2c64936f5c93c10f115a1fef0b763b279"} Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.170010 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab269f92-11bd-4ea1-82a3-a77c15c5bb07","Type":"ContainerStarted","Data":"e19bacf6cf3e243242bf433ee4552047a65e2ea7b1364536e063e65669e24079"} Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.170179 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" containerID="cri-o://9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e" gracePeriod=30 Oct 01 16:07:46 crc kubenswrapper[4688]: I1001 16:07:46.170222 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" containerID="cri-o://f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4" gracePeriod=30 Oct 01 16:07:47 crc kubenswrapper[4688]: I1001 16:07:47.212176 4688 generic.go:334] "Generic (PLEG): container finished" podID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerID="9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e" exitCode=143 Oct 01 16:07:47 crc kubenswrapper[4688]: I1001 16:07:47.212246 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerDied","Data":"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e"} Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.041196 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.116471 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle\") pod \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.116662 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9c8p\" (UniqueName: \"kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p\") pod \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.116742 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data\") pod \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\" (UID: \"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264\") " Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.129677 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p" (OuterVolumeSpecName: "kube-api-access-b9c8p") pod "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" (UID: "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264"). InnerVolumeSpecName "kube-api-access-b9c8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.149557 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" (UID: "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.168761 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data" (OuterVolumeSpecName: "config-data") pod "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" (UID: "8aa3fb86-a78b-4a86-9d94-b3fcc8b48264"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.219733 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.219761 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.219774 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9c8p\" (UniqueName: \"kubernetes.io/projected/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264-kube-api-access-b9c8p\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.222402 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab269f92-11bd-4ea1-82a3-a77c15c5bb07","Type":"ContainerStarted","Data":"a50772559f1f60f1410c7228574b76e033ba52d85d80407920f51e5fb3a3c847"} Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.222570 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.224268 4688 generic.go:334] "Generic (PLEG): container finished" podID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" containerID="5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf" exitCode=0 Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.224299 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264","Type":"ContainerDied","Data":"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf"} Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.224317 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8aa3fb86-a78b-4a86-9d94-b3fcc8b48264","Type":"ContainerDied","Data":"fd431521c1c8e2d27a84910266d9921b8da645c5cb34dd4f929820bba5791864"} Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.224333 4688 scope.go:117] "RemoveContainer" containerID="5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.224592 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.249777 4688 scope.go:117] "RemoveContainer" containerID="5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf" Oct 01 16:07:48 crc kubenswrapper[4688]: E1001 16:07:48.251662 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf\": container with ID starting with 5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf not found: ID does not exist" containerID="5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.251696 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf"} err="failed to get container status \"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf\": rpc error: code = NotFound desc = could not find container \"5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf\": container with ID starting with 5312c09dfa9b18f7734e32423eda17a567f65975de33cbf8064d593db11e6fdf not found: ID does not exist" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.255583 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.065990795 podStartE2EDuration="5.255564461s" podCreationTimestamp="2025-10-01 16:07:43 +0000 UTC" firstStartedPulling="2025-10-01 16:07:44.042030642 +0000 UTC m=+1253.392670604" lastFinishedPulling="2025-10-01 16:07:47.231604308 +0000 UTC m=+1256.582244270" observedRunningTime="2025-10-01 16:07:48.252664368 +0000 UTC m=+1257.603304330" watchObservedRunningTime="2025-10-01 16:07:48.255564461 +0000 UTC m=+1257.606204423" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.271043 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.279876 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.291065 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:48 crc kubenswrapper[4688]: E1001 16:07:48.291439 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8199e7d8-c362-43ae-a734-d287ff9f293d" containerName="nova-manage" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.291451 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8199e7d8-c362-43ae-a734-d287ff9f293d" containerName="nova-manage" Oct 01 16:07:48 crc kubenswrapper[4688]: E1001 16:07:48.291463 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" containerName="nova-scheduler-scheduler" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.291470 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" containerName="nova-scheduler-scheduler" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.291658 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8199e7d8-c362-43ae-a734-d287ff9f293d" containerName="nova-manage" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.291679 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" containerName="nova-scheduler-scheduler" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.292268 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.299423 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.309509 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.423452 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.423599 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bptm9\" (UniqueName: \"kubernetes.io/projected/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-kube-api-access-bptm9\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.423639 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-config-data\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.525660 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.526060 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bptm9\" (UniqueName: \"kubernetes.io/projected/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-kube-api-access-bptm9\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.526264 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-config-data\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.529852 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-config-data\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.532226 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.545096 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bptm9\" (UniqueName: \"kubernetes.io/projected/8d577d2f-a1a6-4d70-8532-86c0bb41f51f-kube-api-access-bptm9\") pod \"nova-scheduler-0\" (UID: \"8d577d2f-a1a6-4d70-8532-86c0bb41f51f\") " pod="openstack/nova-scheduler-0" Oct 01 16:07:48 crc kubenswrapper[4688]: I1001 16:07:48.658901 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.115453 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 16:07:49 crc kubenswrapper[4688]: W1001 16:07:49.119811 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d577d2f_a1a6_4d70_8532_86c0bb41f51f.slice/crio-8f0cbe467254de2a53a28d609e9ead32c9f03b374dbe37d2141bf52591d66b33 WatchSource:0}: Error finding container 8f0cbe467254de2a53a28d609e9ead32c9f03b374dbe37d2141bf52591d66b33: Status 404 returned error can't find the container with id 8f0cbe467254de2a53a28d609e9ead32c9f03b374dbe37d2141bf52591d66b33 Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.264175 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d577d2f-a1a6-4d70-8532-86c0bb41f51f","Type":"ContainerStarted","Data":"8f0cbe467254de2a53a28d609e9ead32c9f03b374dbe37d2141bf52591d66b33"} Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.306950 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:36130->10.217.0.201:8775: read: connection reset by peer" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.307726 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:36114->10.217.0.201:8775: read: connection reset by peer" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.403304 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa3fb86-a78b-4a86-9d94-b3fcc8b48264" path="/var/lib/kubelet/pods/8aa3fb86-a78b-4a86-9d94-b3fcc8b48264/volumes" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.795163 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.967076 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs\") pod \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.967224 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle\") pod \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.967261 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data\") pod \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.967319 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs\") pod \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.967372 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwktn\" (UniqueName: \"kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn\") pod \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\" (UID: \"a1261b09-d2b5-4612-8d4c-d72adccc9aa0\") " Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.968740 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs" (OuterVolumeSpecName: "logs") pod "a1261b09-d2b5-4612-8d4c-d72adccc9aa0" (UID: "a1261b09-d2b5-4612-8d4c-d72adccc9aa0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:49 crc kubenswrapper[4688]: I1001 16:07:49.980818 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn" (OuterVolumeSpecName: "kube-api-access-dwktn") pod "a1261b09-d2b5-4612-8d4c-d72adccc9aa0" (UID: "a1261b09-d2b5-4612-8d4c-d72adccc9aa0"). InnerVolumeSpecName "kube-api-access-dwktn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.000067 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1261b09-d2b5-4612-8d4c-d72adccc9aa0" (UID: "a1261b09-d2b5-4612-8d4c-d72adccc9aa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.013361 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data" (OuterVolumeSpecName: "config-data") pod "a1261b09-d2b5-4612-8d4c-d72adccc9aa0" (UID: "a1261b09-d2b5-4612-8d4c-d72adccc9aa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.035211 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a1261b09-d2b5-4612-8d4c-d72adccc9aa0" (UID: "a1261b09-d2b5-4612-8d4c-d72adccc9aa0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.069228 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.069262 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwktn\" (UniqueName: \"kubernetes.io/projected/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-kube-api-access-dwktn\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.069272 4688 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.069282 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.069291 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1261b09-d2b5-4612-8d4c-d72adccc9aa0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.292006 4688 generic.go:334] "Generic (PLEG): container finished" podID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerID="f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4" exitCode=0 Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.292340 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerDied","Data":"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4"} Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.292375 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1261b09-d2b5-4612-8d4c-d72adccc9aa0","Type":"ContainerDied","Data":"eedfbab440da2cbc043d610a50b5a2f1063933db18f186e6053f2e20fcda145d"} Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.292399 4688 scope.go:117] "RemoveContainer" containerID="f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.292559 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.306152 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d577d2f-a1a6-4d70-8532-86c0bb41f51f","Type":"ContainerStarted","Data":"ed08168ad68c72a14e47c723c190d593cb7c242237ad05c973b9e2f85194c5d2"} Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.336828 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.336814584 podStartE2EDuration="2.336814584s" podCreationTimestamp="2025-10-01 16:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:50.334310853 +0000 UTC m=+1259.684950815" watchObservedRunningTime="2025-10-01 16:07:50.336814584 +0000 UTC m=+1259.687454546" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.371723 4688 scope.go:117] "RemoveContainer" containerID="9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.372562 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.397617 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.404761 4688 scope.go:117] "RemoveContainer" containerID="f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4" Oct 01 16:07:50 crc kubenswrapper[4688]: E1001 16:07:50.405891 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4\": container with ID starting with f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4 not found: ID does not exist" containerID="f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.405935 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4"} err="failed to get container status \"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4\": rpc error: code = NotFound desc = could not find container \"f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4\": container with ID starting with f69259479cc32ce793454880d5c518a5bb65ee5ffbd07d5c8eadae93c3d13ac4 not found: ID does not exist" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.405953 4688 scope.go:117] "RemoveContainer" containerID="9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e" Oct 01 16:07:50 crc kubenswrapper[4688]: E1001 16:07:50.407807 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e\": container with ID starting with 9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e not found: ID does not exist" containerID="9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.407872 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e"} err="failed to get container status \"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e\": rpc error: code = NotFound desc = could not find container \"9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e\": container with ID starting with 9c5404e2f603afc8b91182eb4cf7cdf1696b37690040e0064c130cb47d61ab0e not found: ID does not exist" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.411957 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:50 crc kubenswrapper[4688]: E1001 16:07:50.412315 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.412331 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" Oct 01 16:07:50 crc kubenswrapper[4688]: E1001 16:07:50.412352 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.412360 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.412634 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-log" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.412666 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" containerName="nova-metadata-metadata" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.413619 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.416824 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.416931 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.443007 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.577914 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f50e4d-f039-4902-ab15-03604f701daf-logs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.578002 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.578032 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-config-data\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.578072 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4wx2\" (UniqueName: \"kubernetes.io/projected/c9f50e4d-f039-4902-ab15-03604f701daf-kube-api-access-x4wx2\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.578101 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.679798 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f50e4d-f039-4902-ab15-03604f701daf-logs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.680059 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.680208 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-config-data\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.680337 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4wx2\" (UniqueName: \"kubernetes.io/projected/c9f50e4d-f039-4902-ab15-03604f701daf-kube-api-access-x4wx2\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.680435 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.681051 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f50e4d-f039-4902-ab15-03604f701daf-logs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.690092 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.690359 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.694153 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f50e4d-f039-4902-ab15-03604f701daf-config-data\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.707965 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4wx2\" (UniqueName: \"kubernetes.io/projected/c9f50e4d-f039-4902-ab15-03604f701daf-kube-api-access-x4wx2\") pod \"nova-metadata-0\" (UID: \"c9f50e4d-f039-4902-ab15-03604f701daf\") " pod="openstack/nova-metadata-0" Oct 01 16:07:50 crc kubenswrapper[4688]: I1001 16:07:50.742690 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.063124 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194164 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194564 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194726 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4hcp\" (UniqueName: \"kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194762 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194798 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.194876 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs\") pod \"a062e1e2-1bd3-4543-8513-991de8af5608\" (UID: \"a062e1e2-1bd3-4543-8513-991de8af5608\") " Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.196832 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs" (OuterVolumeSpecName: "logs") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.201167 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp" (OuterVolumeSpecName: "kube-api-access-q4hcp") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "kube-api-access-q4hcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.233234 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.245881 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data" (OuterVolumeSpecName: "config-data") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.256835 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.262208 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a062e1e2-1bd3-4543-8513-991de8af5608" (UID: "a062e1e2-1bd3-4543-8513-991de8af5608"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298052 4688 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298090 4688 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298108 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298122 4688 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a062e1e2-1bd3-4543-8513-991de8af5608-logs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298135 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4hcp\" (UniqueName: \"kubernetes.io/projected/a062e1e2-1bd3-4543-8513-991de8af5608-kube-api-access-q4hcp\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.298148 4688 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a062e1e2-1bd3-4543-8513-991de8af5608-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.321517 4688 generic.go:334] "Generic (PLEG): container finished" podID="a062e1e2-1bd3-4543-8513-991de8af5608" containerID="79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830" exitCode=0 Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.322469 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.324601 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerDied","Data":"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830"} Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.324641 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a062e1e2-1bd3-4543-8513-991de8af5608","Type":"ContainerDied","Data":"2de1718b4bc75300a9b80c2a6e6d33459ea88380a6c87f23307086eb8feb0a20"} Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.324667 4688 scope.go:117] "RemoveContainer" containerID="79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.364047 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.412407 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1261b09-d2b5-4612-8d4c-d72adccc9aa0" path="/var/lib/kubelet/pods/a1261b09-d2b5-4612-8d4c-d72adccc9aa0/volumes" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.507542 4688 scope.go:117] "RemoveContainer" containerID="fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.531027 4688 scope.go:117] "RemoveContainer" containerID="79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830" Oct 01 16:07:51 crc kubenswrapper[4688]: E1001 16:07:51.531652 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830\": container with ID starting with 79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830 not found: ID does not exist" containerID="79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.531785 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830"} err="failed to get container status \"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830\": rpc error: code = NotFound desc = could not find container \"79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830\": container with ID starting with 79092218a2b162d05f448cf12ffb6eca25c552feac6c7ce936685dea49624830 not found: ID does not exist" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.531879 4688 scope.go:117] "RemoveContainer" containerID="fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537" Oct 01 16:07:51 crc kubenswrapper[4688]: E1001 16:07:51.532946 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537\": container with ID starting with fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537 not found: ID does not exist" containerID="fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537" Oct 01 16:07:51 crc kubenswrapper[4688]: I1001 16:07:51.533044 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537"} err="failed to get container status \"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537\": rpc error: code = NotFound desc = could not find container \"fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537\": container with ID starting with fda8ac1c742b29e5af49dc15b49730464685c40fee3c34b00f9e1921889c9537 not found: ID does not exist" Oct 01 16:07:52 crc kubenswrapper[4688]: I1001 16:07:52.335821 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c9f50e4d-f039-4902-ab15-03604f701daf","Type":"ContainerStarted","Data":"a5517133236bcfd4832c7900137cd2f9e443d9b216b4279b1e903f335726951e"} Oct 01 16:07:52 crc kubenswrapper[4688]: I1001 16:07:52.336086 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c9f50e4d-f039-4902-ab15-03604f701daf","Type":"ContainerStarted","Data":"109e0c06162bcdcebc9acd475552ba82e4b0423d1cd69beaeb8b953a4acacf93"} Oct 01 16:07:52 crc kubenswrapper[4688]: I1001 16:07:52.336100 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c9f50e4d-f039-4902-ab15-03604f701daf","Type":"ContainerStarted","Data":"778fe22415a3a40a07e4faeb2a07c9fc702919d4e386adc34ce0751dcb0a9fb5"} Oct 01 16:07:53 crc kubenswrapper[4688]: I1001 16:07:53.659442 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 16:07:55 crc kubenswrapper[4688]: I1001 16:07:55.052716 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:07:55 crc kubenswrapper[4688]: I1001 16:07:55.053444 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:07:55 crc kubenswrapper[4688]: I1001 16:07:55.743958 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:07:55 crc kubenswrapper[4688]: I1001 16:07:55.744009 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 16:07:58 crc kubenswrapper[4688]: I1001 16:07:58.659868 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 16:07:58 crc kubenswrapper[4688]: I1001 16:07:58.692763 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 16:07:58 crc kubenswrapper[4688]: I1001 16:07:58.715021 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=8.715002891 podStartE2EDuration="8.715002891s" podCreationTimestamp="2025-10-01 16:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:07:52.358125513 +0000 UTC m=+1261.708765495" watchObservedRunningTime="2025-10-01 16:07:58.715002891 +0000 UTC m=+1268.065642853" Oct 01 16:07:59 crc kubenswrapper[4688]: I1001 16:07:59.447750 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 16:08:00 crc kubenswrapper[4688]: I1001 16:08:00.743380 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:08:00 crc kubenswrapper[4688]: I1001 16:08:00.743766 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 16:08:01 crc kubenswrapper[4688]: I1001 16:08:01.757837 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c9f50e4d-f039-4902-ab15-03604f701daf" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:08:01 crc kubenswrapper[4688]: I1001 16:08:01.757839 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c9f50e4d-f039-4902-ab15-03604f701daf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:08:10 crc kubenswrapper[4688]: I1001 16:08:10.749572 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:08:10 crc kubenswrapper[4688]: I1001 16:08:10.750146 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 16:08:10 crc kubenswrapper[4688]: I1001 16:08:10.755583 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:08:10 crc kubenswrapper[4688]: I1001 16:08:10.760313 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 16:08:13 crc kubenswrapper[4688]: I1001 16:08:13.670928 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.491623 4688 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poda062e1e2-1bd3-4543-8513-991de8af5608"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poda062e1e2-1bd3-4543-8513-991de8af5608] : Timed out while waiting for systemd to remove kubepods-besteffort-poda062e1e2_1bd3_4543_8513_991de8af5608.slice" Oct 01 16:08:21 crc kubenswrapper[4688]: E1001 16:08:21.492278 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poda062e1e2-1bd3-4543-8513-991de8af5608] : unable to destroy cgroup paths for cgroup [kubepods besteffort poda062e1e2-1bd3-4543-8513-991de8af5608] : Timed out while waiting for systemd to remove kubepods-besteffort-poda062e1e2_1bd3_4543_8513_991de8af5608.slice" pod="openstack/nova-api-0" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.631181 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.692792 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.703008 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.735380 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 16:08:21 crc kubenswrapper[4688]: E1001 16:08:21.736056 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-log" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.736090 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-log" Oct 01 16:08:21 crc kubenswrapper[4688]: E1001 16:08:21.736420 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-api" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.736448 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-api" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.736817 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-log" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.736858 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" containerName="nova-api-api" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.738640 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.746382 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.752862 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.752876 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.753962 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.841464 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-config-data\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.841742 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnw8v\" (UniqueName: \"kubernetes.io/projected/cdb337c8-732c-4123-b47b-212ecb38257d-kube-api-access-tnw8v\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.841831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdb337c8-732c-4123-b47b-212ecb38257d-logs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.841975 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.842093 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.842183 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943335 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943465 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-config-data\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943486 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnw8v\" (UniqueName: \"kubernetes.io/projected/cdb337c8-732c-4123-b47b-212ecb38257d-kube-api-access-tnw8v\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943506 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdb337c8-732c-4123-b47b-212ecb38257d-logs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943560 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.943594 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.944704 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdb337c8-732c-4123-b47b-212ecb38257d-logs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.953196 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.953408 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.954983 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-config-data\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.961543 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdb337c8-732c-4123-b47b-212ecb38257d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:21 crc kubenswrapper[4688]: I1001 16:08:21.969929 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnw8v\" (UniqueName: \"kubernetes.io/projected/cdb337c8-732c-4123-b47b-212ecb38257d-kube-api-access-tnw8v\") pod \"nova-api-0\" (UID: \"cdb337c8-732c-4123-b47b-212ecb38257d\") " pod="openstack/nova-api-0" Oct 01 16:08:22 crc kubenswrapper[4688]: I1001 16:08:22.088097 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 16:08:22 crc kubenswrapper[4688]: I1001 16:08:22.552745 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 16:08:22 crc kubenswrapper[4688]: I1001 16:08:22.640682 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdb337c8-732c-4123-b47b-212ecb38257d","Type":"ContainerStarted","Data":"8a9f2b49a649b8a28172aeb354576c88b2f5c0db6d2027475f1e9ddb5695a399"} Oct 01 16:08:23 crc kubenswrapper[4688]: I1001 16:08:23.397213 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a062e1e2-1bd3-4543-8513-991de8af5608" path="/var/lib/kubelet/pods/a062e1e2-1bd3-4543-8513-991de8af5608/volumes" Oct 01 16:08:23 crc kubenswrapper[4688]: I1001 16:08:23.657285 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdb337c8-732c-4123-b47b-212ecb38257d","Type":"ContainerStarted","Data":"97fd399c2a842ffaddc84afb63cff3b3ffcc5ad2c8c62076da24121075a7f574"} Oct 01 16:08:23 crc kubenswrapper[4688]: I1001 16:08:23.657343 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cdb337c8-732c-4123-b47b-212ecb38257d","Type":"ContainerStarted","Data":"d51b81ef808b912354c577a30edd29172eb87b8a8633dad24a5124aa15b0d75e"} Oct 01 16:08:23 crc kubenswrapper[4688]: I1001 16:08:23.691781 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.6917502559999997 podStartE2EDuration="2.691750256s" podCreationTimestamp="2025-10-01 16:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:08:23.68205836 +0000 UTC m=+1293.032698372" watchObservedRunningTime="2025-10-01 16:08:23.691750256 +0000 UTC m=+1293.042390228" Oct 01 16:08:25 crc kubenswrapper[4688]: I1001 16:08:25.053045 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:08:25 crc kubenswrapper[4688]: I1001 16:08:25.053487 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:08:32 crc kubenswrapper[4688]: I1001 16:08:32.089562 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:08:32 crc kubenswrapper[4688]: I1001 16:08:32.090012 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 16:08:33 crc kubenswrapper[4688]: I1001 16:08:33.106710 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cdb337c8-732c-4123-b47b-212ecb38257d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.210:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:08:33 crc kubenswrapper[4688]: I1001 16:08:33.106710 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cdb337c8-732c-4123-b47b-212ecb38257d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.210:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.095246 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.095898 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.098064 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.098101 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.103895 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:08:42 crc kubenswrapper[4688]: I1001 16:08:42.105249 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 16:08:50 crc kubenswrapper[4688]: I1001 16:08:50.182728 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:08:51 crc kubenswrapper[4688]: I1001 16:08:51.109290 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.053171 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.053773 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.053827 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.054614 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.054672 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c" gracePeriod=600 Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.401191 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="rabbitmq" containerID="cri-o://54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd" gracePeriod=604795 Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.980791 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c" exitCode=0 Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.980986 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c"} Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.981140 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7"} Oct 01 16:08:55 crc kubenswrapper[4688]: I1001 16:08:55.981164 4688 scope.go:117] "RemoveContainer" containerID="aa1d24349df20faa30a92dfa493f4802fd2f9c2620bfb173aa37e4d116500a18" Oct 01 16:08:56 crc kubenswrapper[4688]: I1001 16:08:56.080332 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="rabbitmq" containerID="cri-o://9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49" gracePeriod=604796 Oct 01 16:08:57 crc kubenswrapper[4688]: I1001 16:08:57.140362 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 01 16:08:57 crc kubenswrapper[4688]: I1001 16:08:57.517374 4688 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.052364 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.052353 4688 generic.go:334] "Generic (PLEG): container finished" podID="0372e042-2565-4498-8884-452f6e0bd0f1" containerID="54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd" exitCode=0 Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.052386 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerDied","Data":"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd"} Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.053021 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0372e042-2565-4498-8884-452f6e0bd0f1","Type":"ContainerDied","Data":"0c08901b48369fcd0d0be8e38b86f17709b966518249b6a7214d13123faab7e2"} Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.053058 4688 scope.go:117] "RemoveContainer" containerID="54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.091476 4688 scope.go:117] "RemoveContainer" containerID="b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.153845 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.153974 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154012 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154051 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqwxb\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154078 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154120 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154165 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154217 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154263 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154305 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.154334 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie\") pod \"0372e042-2565-4498-8884-452f6e0bd0f1\" (UID: \"0372e042-2565-4498-8884-452f6e0bd0f1\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.160989 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.174123 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.176311 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.179961 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.185570 4688 scope.go:117] "RemoveContainer" containerID="54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd" Oct 01 16:09:02 crc kubenswrapper[4688]: E1001 16:09:02.185978 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd\": container with ID starting with 54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd not found: ID does not exist" containerID="54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.186011 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd"} err="failed to get container status \"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd\": rpc error: code = NotFound desc = could not find container \"54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd\": container with ID starting with 54c2e2d595d7817fccc48fb84f1a99abf50a63b8491171376196751f0a15ddbd not found: ID does not exist" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.186032 4688 scope.go:117] "RemoveContainer" containerID="b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d" Oct 01 16:09:02 crc kubenswrapper[4688]: E1001 16:09:02.186213 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d\": container with ID starting with b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d not found: ID does not exist" containerID="b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.186233 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d"} err="failed to get container status \"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d\": rpc error: code = NotFound desc = could not find container \"b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d\": container with ID starting with b55b9135521970589e3f5ba78a330ef017909c6222e7bba8e3ebd14d71be440d not found: ID does not exist" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.186569 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb" (OuterVolumeSpecName: "kube-api-access-gqwxb") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "kube-api-access-gqwxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.199954 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.202420 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info" (OuterVolumeSpecName: "pod-info") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.208376 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.261870 4688 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262111 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262128 4688 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0372e042-2565-4498-8884-452f6e0bd0f1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262163 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262177 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqwxb\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-kube-api-access-gqwxb\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262188 4688 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0372e042-2565-4498-8884-452f6e0bd0f1-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262201 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.262212 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.278136 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf" (OuterVolumeSpecName: "server-conf") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.284276 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data" (OuterVolumeSpecName: "config-data") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.295736 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.346946 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0372e042-2565-4498-8884-452f6e0bd0f1" (UID: "0372e042-2565-4498-8884-452f6e0bd0f1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.363814 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.363845 4688 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0372e042-2565-4498-8884-452f6e0bd0f1-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.363856 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0372e042-2565-4498-8884-452f6e0bd0f1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.363867 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.683622 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770439 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770627 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770654 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770751 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770771 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770791 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770851 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndqkd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770900 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770945 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.770963 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.771016 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf\") pod \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\" (UID: \"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c\") " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.774603 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.775655 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.778496 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.780091 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.784258 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info" (OuterVolumeSpecName: "pod-info") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.794983 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.795053 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.815246 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd" (OuterVolumeSpecName: "kube-api-access-ndqkd") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "kube-api-access-ndqkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.832244 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data" (OuterVolumeSpecName: "config-data") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873457 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873494 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873508 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873517 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873536 4688 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873545 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndqkd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-kube-api-access-ndqkd\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873553 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873563 4688 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.873571 4688 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.915529 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf" (OuterVolumeSpecName: "server-conf") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.978746 4688 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:02 crc kubenswrapper[4688]: I1001 16:09:02.984156 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.045375 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" (UID: "51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.068078 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.075898 4688 generic.go:334] "Generic (PLEG): container finished" podID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerID="9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49" exitCode=0 Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.075934 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerDied","Data":"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49"} Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.075957 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c","Type":"ContainerDied","Data":"146617fef1bc46b28b0ee849ca910185153616c26c899173704406e6f5fb6f24"} Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.075974 4688 scope.go:117] "RemoveContainer" containerID="9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.076181 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.086167 4688 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.086477 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.131979 4688 scope.go:117] "RemoveContainer" containerID="f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.139774 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.173746 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.200512 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.201025 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201046 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.201066 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201073 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.201087 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="setup-container" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201093 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="setup-container" Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.201105 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="setup-container" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201112 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="setup-container" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201298 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.201313 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" containerName="rabbitmq" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.202262 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.202489 4688 scope.go:117] "RemoveContainer" containerID="9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49" Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.203350 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49\": container with ID starting with 9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49 not found: ID does not exist" containerID="9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.203372 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49"} err="failed to get container status \"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49\": rpc error: code = NotFound desc = could not find container \"9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49\": container with ID starting with 9095aaf529d78e89b2a665506c3e9d5868878ac010f0bd5825a725a50ba25f49 not found: ID does not exist" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.203392 4688 scope.go:117] "RemoveContainer" containerID="f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567" Oct 01 16:09:03 crc kubenswrapper[4688]: E1001 16:09:03.203830 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567\": container with ID starting with f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567 not found: ID does not exist" containerID="f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.203845 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567"} err="failed to get container status \"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567\": rpc error: code = NotFound desc = could not find container \"f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567\": container with ID starting with f82a45b830e46e43e476043e4098ddd6b6ad7c5c1b4b3c0af02c3f1a90303567 not found: ID does not exist" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.207305 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.207473 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.207680 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.207810 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.207932 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.208142 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.212555 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.212942 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mdzlh" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.237642 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.267318 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.275832 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.277515 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.283215 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296308 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tmrnr" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296499 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296698 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296868 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296934 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.296882 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.299127 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.301973 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302041 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302122 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rshwm\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-kube-api-access-rshwm\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302186 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302249 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d1876945-781a-41af-a495-bee2af872b98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302289 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302314 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302344 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302367 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302387 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d1876945-781a-41af-a495-bee2af872b98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.302491 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.402336 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0372e042-2565-4498-8884-452f6e0bd0f1" path="/var/lib/kubelet/pods/0372e042-2565-4498-8884-452f6e0bd0f1/volumes" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404551 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404590 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404628 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404650 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rshwm\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-kube-api-access-rshwm\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404681 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404710 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404740 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d1876945-781a-41af-a495-bee2af872b98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404760 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404775 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404805 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404823 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404851 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404872 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404886 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d1876945-781a-41af-a495-bee2af872b98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404912 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404927 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404947 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94mh\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-kube-api-access-g94mh\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404982 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.404998 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.405016 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.405038 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.405068 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.405254 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c" path="/var/lib/kubelet/pods/51ba652f-dca6-4c1b-9d78-1ac10f1bcb3c/volumes" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.406193 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.408301 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.408945 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.409165 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.409650 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.411648 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1876945-781a-41af-a495-bee2af872b98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.412226 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d1876945-781a-41af-a495-bee2af872b98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.412635 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d1876945-781a-41af-a495-bee2af872b98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.413197 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.416591 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.429665 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rshwm\" (UniqueName: \"kubernetes.io/projected/d1876945-781a-41af-a495-bee2af872b98-kube-api-access-rshwm\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.462422 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d1876945-781a-41af-a495-bee2af872b98\") " pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506443 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506569 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506602 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506669 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506729 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506787 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506811 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506868 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506893 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506917 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94mh\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-kube-api-access-g94mh\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.506966 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.507666 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.508128 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.508497 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.508672 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.508864 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.509600 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.514549 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.515338 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.515569 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.515846 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.527970 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94mh\" (UniqueName: \"kubernetes.io/projected/e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb-kube-api-access-g94mh\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.537503 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.570883 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:03 crc kubenswrapper[4688]: I1001 16:09:03.608101 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:04 crc kubenswrapper[4688]: I1001 16:09:04.053164 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 16:09:04 crc kubenswrapper[4688]: I1001 16:09:04.099454 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d1876945-781a-41af-a495-bee2af872b98","Type":"ContainerStarted","Data":"dd35d72f980cbb5e8970a29b867c1d74c36bd3d20bf89531704f59e40640e021"} Oct 01 16:09:04 crc kubenswrapper[4688]: I1001 16:09:04.169298 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 16:09:05 crc kubenswrapper[4688]: I1001 16:09:05.109047 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb","Type":"ContainerStarted","Data":"e1693781d15a46e172d88cba0a53e245c16c3650da6369c5b445d649733bb1c3"} Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.118756 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb","Type":"ContainerStarted","Data":"2cd58e1886a7b4effd5afd8a88ce1d20bc98a45eb2390c8f045085ef00087458"} Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.120836 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d1876945-781a-41af-a495-bee2af872b98","Type":"ContainerStarted","Data":"0799a6915e00ab8210296a573620e08d6eb0c48334f7e836b7b0e254dd2926bd"} Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.337633 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.339592 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.348079 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.359724 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361662 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361742 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361784 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361812 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361829 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwrt\" (UniqueName: \"kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361887 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.361915 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464297 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464408 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464459 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464480 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwrt\" (UniqueName: \"kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464629 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464663 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.464787 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.465124 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.465254 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.465628 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.465991 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.466107 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.466209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.498948 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwrt\" (UniqueName: \"kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt\") pod \"dnsmasq-dns-d558885bc-zrthm\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.664456 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:06 crc kubenswrapper[4688]: I1001 16:09:06.977036 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:07 crc kubenswrapper[4688]: I1001 16:09:07.135891 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-zrthm" event={"ID":"5568ba3d-aca1-4916-903d-a8307b963e0b","Type":"ContainerStarted","Data":"80dce729e56316ff5e8b084a1ea47dabbee2aad0947fad6079a59816082c72f1"} Oct 01 16:09:08 crc kubenswrapper[4688]: I1001 16:09:08.148804 4688 generic.go:334] "Generic (PLEG): container finished" podID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerID="969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0" exitCode=0 Oct 01 16:09:08 crc kubenswrapper[4688]: I1001 16:09:08.148889 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-zrthm" event={"ID":"5568ba3d-aca1-4916-903d-a8307b963e0b","Type":"ContainerDied","Data":"969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0"} Oct 01 16:09:09 crc kubenswrapper[4688]: I1001 16:09:09.160713 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-zrthm" event={"ID":"5568ba3d-aca1-4916-903d-a8307b963e0b","Type":"ContainerStarted","Data":"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3"} Oct 01 16:09:09 crc kubenswrapper[4688]: I1001 16:09:09.161109 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:09 crc kubenswrapper[4688]: I1001 16:09:09.183601 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-zrthm" podStartSLOduration=3.183585454 podStartE2EDuration="3.183585454s" podCreationTimestamp="2025-10-01 16:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:09:09.178380395 +0000 UTC m=+1338.529020367" watchObservedRunningTime="2025-10-01 16:09:09.183585454 +0000 UTC m=+1338.534225416" Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.665801 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.746940 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.747279 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="dnsmasq-dns" containerID="cri-o://f4a78be3c595360ad8f3085e62a0ea1c42e1ad7f88737f5a98b22f4d4a7b7fc3" gracePeriod=10 Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.932655 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-vhf77"] Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.934493 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:16 crc kubenswrapper[4688]: I1001 16:09:16.976350 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-vhf77"] Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.006883 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.006940 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-config\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.006983 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.007050 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.007088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfnvn\" (UniqueName: \"kubernetes.io/projected/3b618166-5602-4785-8893-eb7b86fbe39a-kube-api-access-zfnvn\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.007117 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.007164 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.108824 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.108902 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.108950 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-config\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.108986 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.109084 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.109141 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfnvn\" (UniqueName: \"kubernetes.io/projected/3b618166-5602-4785-8893-eb7b86fbe39a-kube-api-access-zfnvn\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.109172 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.109960 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-nb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.110222 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-swift-storage-0\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.110248 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-config\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.110615 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-dns-svc\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.110986 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.118881 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b618166-5602-4785-8893-eb7b86fbe39a-ovsdbserver-sb\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.127945 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfnvn\" (UniqueName: \"kubernetes.io/projected/3b618166-5602-4785-8893-eb7b86fbe39a-kube-api-access-zfnvn\") pod \"dnsmasq-dns-6b6dc74c5-vhf77\" (UID: \"3b618166-5602-4785-8893-eb7b86fbe39a\") " pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.236559 4688 generic.go:334] "Generic (PLEG): container finished" podID="ea25de60-738d-40ce-bc03-37155b616e1e" containerID="f4a78be3c595360ad8f3085e62a0ea1c42e1ad7f88737f5a98b22f4d4a7b7fc3" exitCode=0 Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.236641 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" event={"ID":"ea25de60-738d-40ce-bc03-37155b616e1e","Type":"ContainerDied","Data":"f4a78be3c595360ad8f3085e62a0ea1c42e1ad7f88737f5a98b22f4d4a7b7fc3"} Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.289938 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.426839 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516510 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516570 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsc5d\" (UniqueName: \"kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516699 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516749 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516792 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.516818 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb\") pod \"ea25de60-738d-40ce-bc03-37155b616e1e\" (UID: \"ea25de60-738d-40ce-bc03-37155b616e1e\") " Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.548181 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d" (OuterVolumeSpecName: "kube-api-access-rsc5d") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "kube-api-access-rsc5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.625565 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsc5d\" (UniqueName: \"kubernetes.io/projected/ea25de60-738d-40ce-bc03-37155b616e1e-kube-api-access-rsc5d\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.631601 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.632133 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config" (OuterVolumeSpecName: "config") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.655542 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.660897 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.685430 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea25de60-738d-40ce-bc03-37155b616e1e" (UID: "ea25de60-738d-40ce-bc03-37155b616e1e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.727775 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.727825 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.727839 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.727851 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.727862 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea25de60-738d-40ce-bc03-37155b616e1e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:17 crc kubenswrapper[4688]: I1001 16:09:17.943974 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b6dc74c5-vhf77"] Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.251561 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" event={"ID":"ea25de60-738d-40ce-bc03-37155b616e1e","Type":"ContainerDied","Data":"dc25c7734b62312080358dc57a63bb65ee919ed7cfaead1e2641619c93f5bd2d"} Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.252177 4688 scope.go:117] "RemoveContainer" containerID="f4a78be3c595360ad8f3085e62a0ea1c42e1ad7f88737f5a98b22f4d4a7b7fc3" Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.251807 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-8vh5d" Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.253423 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" event={"ID":"3b618166-5602-4785-8893-eb7b86fbe39a","Type":"ContainerStarted","Data":"924d93de94060cac2f4e334ff6cc261b53f06fd7e39a34a0b5d56794847e5442"} Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.284240 4688 scope.go:117] "RemoveContainer" containerID="634215ffb6afedf22db1c31fe24c482376cbd5dd700f4f6ee15d5d9c836ac1e2" Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.287699 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:09:18 crc kubenswrapper[4688]: I1001 16:09:18.294347 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-8vh5d"] Oct 01 16:09:19 crc kubenswrapper[4688]: I1001 16:09:19.275843 4688 generic.go:334] "Generic (PLEG): container finished" podID="3b618166-5602-4785-8893-eb7b86fbe39a" containerID="3cee3eed047ae145e7017d82c42185257815032023e6f7eb0ce738bbc2eb07be" exitCode=0 Oct 01 16:09:19 crc kubenswrapper[4688]: I1001 16:09:19.275895 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" event={"ID":"3b618166-5602-4785-8893-eb7b86fbe39a","Type":"ContainerDied","Data":"3cee3eed047ae145e7017d82c42185257815032023e6f7eb0ce738bbc2eb07be"} Oct 01 16:09:19 crc kubenswrapper[4688]: I1001 16:09:19.395485 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" path="/var/lib/kubelet/pods/ea25de60-738d-40ce-bc03-37155b616e1e/volumes" Oct 01 16:09:20 crc kubenswrapper[4688]: I1001 16:09:20.286509 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" event={"ID":"3b618166-5602-4785-8893-eb7b86fbe39a","Type":"ContainerStarted","Data":"6bda98fc56153d95d1d1cbbe4b09055111af0b8472c8f84ebcedd86d7c81b77f"} Oct 01 16:09:20 crc kubenswrapper[4688]: I1001 16:09:20.286947 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:20 crc kubenswrapper[4688]: I1001 16:09:20.308150 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" podStartSLOduration=4.3081301530000005 podStartE2EDuration="4.308130153s" podCreationTimestamp="2025-10-01 16:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:09:20.306045454 +0000 UTC m=+1349.656685416" watchObservedRunningTime="2025-10-01 16:09:20.308130153 +0000 UTC m=+1349.658770115" Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.291726 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b6dc74c5-vhf77" Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.376122 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.376491 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-zrthm" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="dnsmasq-dns" containerID="cri-o://b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3" gracePeriod=10 Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.888761 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.946727 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.946823 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.946873 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgwrt\" (UniqueName: \"kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.946926 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.946949 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.947000 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.947080 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc\") pod \"5568ba3d-aca1-4916-903d-a8307b963e0b\" (UID: \"5568ba3d-aca1-4916-903d-a8307b963e0b\") " Oct 01 16:09:27 crc kubenswrapper[4688]: I1001 16:09:27.959134 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt" (OuterVolumeSpecName: "kube-api-access-jgwrt") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "kube-api-access-jgwrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.012991 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.028045 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.028782 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.032196 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.054769 4688 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.054803 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.054813 4688 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.054824 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgwrt\" (UniqueName: \"kubernetes.io/projected/5568ba3d-aca1-4916-903d-a8307b963e0b-kube-api-access-jgwrt\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.054840 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.058962 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.059019 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config" (OuterVolumeSpecName: "config") pod "5568ba3d-aca1-4916-903d-a8307b963e0b" (UID: "5568ba3d-aca1-4916-903d-a8307b963e0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.156420 4688 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.156455 4688 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5568ba3d-aca1-4916-903d-a8307b963e0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.376744 4688 generic.go:334] "Generic (PLEG): container finished" podID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerID="b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3" exitCode=0 Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.376795 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-zrthm" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.376789 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-zrthm" event={"ID":"5568ba3d-aca1-4916-903d-a8307b963e0b","Type":"ContainerDied","Data":"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3"} Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.377645 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-zrthm" event={"ID":"5568ba3d-aca1-4916-903d-a8307b963e0b","Type":"ContainerDied","Data":"80dce729e56316ff5e8b084a1ea47dabbee2aad0947fad6079a59816082c72f1"} Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.377674 4688 scope.go:117] "RemoveContainer" containerID="b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.398626 4688 scope.go:117] "RemoveContainer" containerID="969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.420807 4688 scope.go:117] "RemoveContainer" containerID="b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3" Oct 01 16:09:28 crc kubenswrapper[4688]: E1001 16:09:28.421251 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3\": container with ID starting with b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3 not found: ID does not exist" containerID="b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.421278 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3"} err="failed to get container status \"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3\": rpc error: code = NotFound desc = could not find container \"b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3\": container with ID starting with b06b2fca295f092152b5af755e75db308881332a958239f0c1af2dece4e2aee3 not found: ID does not exist" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.421297 4688 scope.go:117] "RemoveContainer" containerID="969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0" Oct 01 16:09:28 crc kubenswrapper[4688]: E1001 16:09:28.421658 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0\": container with ID starting with 969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0 not found: ID does not exist" containerID="969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.421679 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0"} err="failed to get container status \"969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0\": rpc error: code = NotFound desc = could not find container \"969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0\": container with ID starting with 969146df62c6086180e054a1d9f533a80c21a9b7e0c32703b56a3be2fabc7dd0 not found: ID does not exist" Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.431469 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:28 crc kubenswrapper[4688]: I1001 16:09:28.439461 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-zrthm"] Oct 01 16:09:29 crc kubenswrapper[4688]: I1001 16:09:29.390917 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" path="/var/lib/kubelet/pods/5568ba3d-aca1-4916-903d-a8307b963e0b/volumes" Oct 01 16:09:38 crc kubenswrapper[4688]: I1001 16:09:38.467695 4688 generic.go:334] "Generic (PLEG): container finished" podID="d1876945-781a-41af-a495-bee2af872b98" containerID="0799a6915e00ab8210296a573620e08d6eb0c48334f7e836b7b0e254dd2926bd" exitCode=0 Oct 01 16:09:38 crc kubenswrapper[4688]: I1001 16:09:38.467730 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d1876945-781a-41af-a495-bee2af872b98","Type":"ContainerDied","Data":"0799a6915e00ab8210296a573620e08d6eb0c48334f7e836b7b0e254dd2926bd"} Oct 01 16:09:38 crc kubenswrapper[4688]: I1001 16:09:38.472352 4688 generic.go:334] "Generic (PLEG): container finished" podID="e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb" containerID="2cd58e1886a7b4effd5afd8a88ce1d20bc98a45eb2390c8f045085ef00087458" exitCode=0 Oct 01 16:09:38 crc kubenswrapper[4688]: I1001 16:09:38.472386 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb","Type":"ContainerDied","Data":"2cd58e1886a7b4effd5afd8a88ce1d20bc98a45eb2390c8f045085ef00087458"} Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.484287 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb","Type":"ContainerStarted","Data":"b88d705328bb041f3850048546a33a0459363865048e8a1b6deaa0eccad67cd7"} Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.485202 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.487016 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d1876945-781a-41af-a495-bee2af872b98","Type":"ContainerStarted","Data":"6b6b9a51d15e2af741264af22856458c035c525bf881c3f697281c661ee544e0"} Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.487288 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.516052 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.51603048 podStartE2EDuration="36.51603048s" podCreationTimestamp="2025-10-01 16:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:09:39.513868788 +0000 UTC m=+1368.864508750" watchObservedRunningTime="2025-10-01 16:09:39.51603048 +0000 UTC m=+1368.866670442" Oct 01 16:09:39 crc kubenswrapper[4688]: I1001 16:09:39.542258 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.542242198 podStartE2EDuration="36.542242198s" podCreationTimestamp="2025-10-01 16:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:09:39.538249754 +0000 UTC m=+1368.888889726" watchObservedRunningTime="2025-10-01 16:09:39.542242198 +0000 UTC m=+1368.892882160" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.938309 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc"] Oct 01 16:09:45 crc kubenswrapper[4688]: E1001 16:09:45.939217 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="init" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939233 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="init" Oct 01 16:09:45 crc kubenswrapper[4688]: E1001 16:09:45.939251 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="init" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939258 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="init" Oct 01 16:09:45 crc kubenswrapper[4688]: E1001 16:09:45.939272 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939282 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: E1001 16:09:45.939298 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939306 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939481 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea25de60-738d-40ce-bc03-37155b616e1e" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.939507 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="5568ba3d-aca1-4916-903d-a8307b963e0b" containerName="dnsmasq-dns" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.944390 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.947565 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.948255 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.948478 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.948630 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:09:45 crc kubenswrapper[4688]: I1001 16:09:45.972954 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc"] Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.008495 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c565j\" (UniqueName: \"kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.008598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.008644 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.008700 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.110048 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c565j\" (UniqueName: \"kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.110128 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.110173 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.110223 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.115070 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.115360 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.115572 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.134013 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c565j\" (UniqueName: \"kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:46 crc kubenswrapper[4688]: I1001 16:09:46.265556 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:09:47 crc kubenswrapper[4688]: I1001 16:09:47.053546 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc"] Oct 01 16:09:47 crc kubenswrapper[4688]: I1001 16:09:47.554133 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" event={"ID":"d8469466-3ce0-4d29-8d96-6cce137617fa","Type":"ContainerStarted","Data":"ad6ef90f8888237517d641d280974a9d620dc5668a50f7008243046522889eeb"} Oct 01 16:09:53 crc kubenswrapper[4688]: I1001 16:09:53.541793 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 16:09:53 crc kubenswrapper[4688]: I1001 16:09:53.611728 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 16:09:56 crc kubenswrapper[4688]: I1001 16:09:56.824385 4688 scope.go:117] "RemoveContainer" containerID="56373d3e4abb22326b033e7c777f0942a0224d5d62b4c5463e1734dd83e33749" Oct 01 16:09:59 crc kubenswrapper[4688]: I1001 16:09:59.055607 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:09:59 crc kubenswrapper[4688]: I1001 16:09:59.725949 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" event={"ID":"d8469466-3ce0-4d29-8d96-6cce137617fa","Type":"ContainerStarted","Data":"2255c356a2aaaa048199c0845983aa892316e368a28468bc2f706d0fd08feffb"} Oct 01 16:09:59 crc kubenswrapper[4688]: I1001 16:09:59.747780 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" podStartSLOduration=2.766529879 podStartE2EDuration="14.747758215s" podCreationTimestamp="2025-10-01 16:09:45 +0000 UTC" firstStartedPulling="2025-10-01 16:09:47.071561459 +0000 UTC m=+1376.422201421" lastFinishedPulling="2025-10-01 16:09:59.052789785 +0000 UTC m=+1388.403429757" observedRunningTime="2025-10-01 16:09:59.73845283 +0000 UTC m=+1389.089092802" watchObservedRunningTime="2025-10-01 16:09:59.747758215 +0000 UTC m=+1389.098398177" Oct 01 16:10:12 crc kubenswrapper[4688]: I1001 16:10:12.884114 4688 generic.go:334] "Generic (PLEG): container finished" podID="d8469466-3ce0-4d29-8d96-6cce137617fa" containerID="2255c356a2aaaa048199c0845983aa892316e368a28468bc2f706d0fd08feffb" exitCode=0 Oct 01 16:10:12 crc kubenswrapper[4688]: I1001 16:10:12.884235 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" event={"ID":"d8469466-3ce0-4d29-8d96-6cce137617fa","Type":"ContainerDied","Data":"2255c356a2aaaa048199c0845983aa892316e368a28468bc2f706d0fd08feffb"} Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.396719 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.567434 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory\") pod \"d8469466-3ce0-4d29-8d96-6cce137617fa\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.567645 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c565j\" (UniqueName: \"kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j\") pod \"d8469466-3ce0-4d29-8d96-6cce137617fa\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.567754 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key\") pod \"d8469466-3ce0-4d29-8d96-6cce137617fa\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.567778 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle\") pod \"d8469466-3ce0-4d29-8d96-6cce137617fa\" (UID: \"d8469466-3ce0-4d29-8d96-6cce137617fa\") " Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.573716 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j" (OuterVolumeSpecName: "kube-api-access-c565j") pod "d8469466-3ce0-4d29-8d96-6cce137617fa" (UID: "d8469466-3ce0-4d29-8d96-6cce137617fa"). InnerVolumeSpecName "kube-api-access-c565j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.574561 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d8469466-3ce0-4d29-8d96-6cce137617fa" (UID: "d8469466-3ce0-4d29-8d96-6cce137617fa"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.599810 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory" (OuterVolumeSpecName: "inventory") pod "d8469466-3ce0-4d29-8d96-6cce137617fa" (UID: "d8469466-3ce0-4d29-8d96-6cce137617fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.602735 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8469466-3ce0-4d29-8d96-6cce137617fa" (UID: "d8469466-3ce0-4d29-8d96-6cce137617fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.670358 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.670512 4688 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.670597 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8469466-3ce0-4d29-8d96-6cce137617fa-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.670677 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c565j\" (UniqueName: \"kubernetes.io/projected/d8469466-3ce0-4d29-8d96-6cce137617fa-kube-api-access-c565j\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.906850 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" event={"ID":"d8469466-3ce0-4d29-8d96-6cce137617fa","Type":"ContainerDied","Data":"ad6ef90f8888237517d641d280974a9d620dc5668a50f7008243046522889eeb"} Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.906895 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad6ef90f8888237517d641d280974a9d620dc5668a50f7008243046522889eeb" Oct 01 16:10:14 crc kubenswrapper[4688]: I1001 16:10:14.906924 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.012864 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj"] Oct 01 16:10:15 crc kubenswrapper[4688]: E1001 16:10:15.013365 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8469466-3ce0-4d29-8d96-6cce137617fa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.013384 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8469466-3ce0-4d29-8d96-6cce137617fa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.013658 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8469466-3ce0-4d29-8d96-6cce137617fa" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.014412 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.018023 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.018359 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.018817 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.019463 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.027262 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj"] Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.179740 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.180130 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqkdg\" (UniqueName: \"kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.180328 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.282210 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.282326 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.282429 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqkdg\" (UniqueName: \"kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.287847 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.288054 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.314666 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqkdg\" (UniqueName: \"kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7tdxj\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.351397 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:15 crc kubenswrapper[4688]: I1001 16:10:15.925053 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj"] Oct 01 16:10:16 crc kubenswrapper[4688]: I1001 16:10:16.929801 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" event={"ID":"1d630ed1-0ec0-4b50-aef4-b8a75aa67270","Type":"ContainerStarted","Data":"8a2eb71d8917412da4a7aa80109f26509ebd5f62093db62330b2274d4970a48d"} Oct 01 16:10:16 crc kubenswrapper[4688]: I1001 16:10:16.930083 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" event={"ID":"1d630ed1-0ec0-4b50-aef4-b8a75aa67270","Type":"ContainerStarted","Data":"76bb9209cb5b731ea1f89f59978c44a1bb817c9f6bf39c3b63ae39f817489757"} Oct 01 16:10:16 crc kubenswrapper[4688]: I1001 16:10:16.954620 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" podStartSLOduration=2.5660979040000003 podStartE2EDuration="2.954578804s" podCreationTimestamp="2025-10-01 16:10:14 +0000 UTC" firstStartedPulling="2025-10-01 16:10:15.930848679 +0000 UTC m=+1405.281488661" lastFinishedPulling="2025-10-01 16:10:16.319329579 +0000 UTC m=+1405.669969561" observedRunningTime="2025-10-01 16:10:16.950354733 +0000 UTC m=+1406.300994715" watchObservedRunningTime="2025-10-01 16:10:16.954578804 +0000 UTC m=+1406.305218776" Oct 01 16:10:19 crc kubenswrapper[4688]: I1001 16:10:19.959359 4688 generic.go:334] "Generic (PLEG): container finished" podID="1d630ed1-0ec0-4b50-aef4-b8a75aa67270" containerID="8a2eb71d8917412da4a7aa80109f26509ebd5f62093db62330b2274d4970a48d" exitCode=0 Oct 01 16:10:19 crc kubenswrapper[4688]: I1001 16:10:19.959449 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" event={"ID":"1d630ed1-0ec0-4b50-aef4-b8a75aa67270","Type":"ContainerDied","Data":"8a2eb71d8917412da4a7aa80109f26509ebd5f62093db62330b2274d4970a48d"} Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.434823 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.611684 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqkdg\" (UniqueName: \"kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg\") pod \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.612590 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory\") pod \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.612921 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key\") pod \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\" (UID: \"1d630ed1-0ec0-4b50-aef4-b8a75aa67270\") " Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.616484 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg" (OuterVolumeSpecName: "kube-api-access-nqkdg") pod "1d630ed1-0ec0-4b50-aef4-b8a75aa67270" (UID: "1d630ed1-0ec0-4b50-aef4-b8a75aa67270"). InnerVolumeSpecName "kube-api-access-nqkdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.639411 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory" (OuterVolumeSpecName: "inventory") pod "1d630ed1-0ec0-4b50-aef4-b8a75aa67270" (UID: "1d630ed1-0ec0-4b50-aef4-b8a75aa67270"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.647852 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d630ed1-0ec0-4b50-aef4-b8a75aa67270" (UID: "1d630ed1-0ec0-4b50-aef4-b8a75aa67270"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.715544 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.715569 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqkdg\" (UniqueName: \"kubernetes.io/projected/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-kube-api-access-nqkdg\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.715580 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d630ed1-0ec0-4b50-aef4-b8a75aa67270-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.983128 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" event={"ID":"1d630ed1-0ec0-4b50-aef4-b8a75aa67270","Type":"ContainerDied","Data":"76bb9209cb5b731ea1f89f59978c44a1bb817c9f6bf39c3b63ae39f817489757"} Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.983172 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76bb9209cb5b731ea1f89f59978c44a1bb817c9f6bf39c3b63ae39f817489757" Oct 01 16:10:21 crc kubenswrapper[4688]: I1001 16:10:21.983179 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7tdxj" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.106229 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825"] Oct 01 16:10:22 crc kubenswrapper[4688]: E1001 16:10:22.106715 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d630ed1-0ec0-4b50-aef4-b8a75aa67270" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.106739 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d630ed1-0ec0-4b50-aef4-b8a75aa67270" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.107405 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d630ed1-0ec0-4b50-aef4-b8a75aa67270" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.109163 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.113411 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.113692 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.113804 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.113937 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.135726 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.135773 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfms5\" (UniqueName: \"kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.135820 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.135894 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.144129 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825"] Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.236831 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.236881 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfms5\" (UniqueName: \"kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.236921 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.236969 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.241338 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.250720 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.252094 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.261076 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfms5\" (UniqueName: \"kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bk825\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:22 crc kubenswrapper[4688]: I1001 16:10:22.443128 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:10:23 crc kubenswrapper[4688]: I1001 16:10:23.001771 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825"] Oct 01 16:10:24 crc kubenswrapper[4688]: I1001 16:10:24.006250 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" event={"ID":"fbd7ce66-5861-4a86-a892-ce6493baf848","Type":"ContainerStarted","Data":"485b7601e0bf7f5860dca83c9434747bfffa78a596ffbf41abb01500e1affbfc"} Oct 01 16:10:24 crc kubenswrapper[4688]: I1001 16:10:24.006506 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" event={"ID":"fbd7ce66-5861-4a86-a892-ce6493baf848","Type":"ContainerStarted","Data":"caf3344f3173d26242eb545f7ec03cc0d582ecf589cd1d6bae6fc6aa91b09457"} Oct 01 16:10:24 crc kubenswrapper[4688]: I1001 16:10:24.037394 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" podStartSLOduration=1.6093126359999999 podStartE2EDuration="2.037375367s" podCreationTimestamp="2025-10-01 16:10:22 +0000 UTC" firstStartedPulling="2025-10-01 16:10:22.997155841 +0000 UTC m=+1412.347795803" lastFinishedPulling="2025-10-01 16:10:23.425218572 +0000 UTC m=+1412.775858534" observedRunningTime="2025-10-01 16:10:24.030619624 +0000 UTC m=+1413.381259596" watchObservedRunningTime="2025-10-01 16:10:24.037375367 +0000 UTC m=+1413.388015339" Oct 01 16:10:55 crc kubenswrapper[4688]: I1001 16:10:55.053257 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:10:55 crc kubenswrapper[4688]: I1001 16:10:55.053989 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:10:59 crc kubenswrapper[4688]: I1001 16:10:59.098926 4688 scope.go:117] "RemoveContainer" containerID="2b7f7e054af4f411521fe37a0501664525292ea06e93ed75734516903533122c" Oct 01 16:10:59 crc kubenswrapper[4688]: I1001 16:10:59.122659 4688 scope.go:117] "RemoveContainer" containerID="77433936d791e20959e176cedf7427bd592cbdef3cc6723d0c7b18814b89c22d" Oct 01 16:10:59 crc kubenswrapper[4688]: I1001 16:10:59.144468 4688 scope.go:117] "RemoveContainer" containerID="87e25ddf63b34ce84c3833804485adea9b8b919ad25c39fd02707d4797968acb" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.239519 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.242092 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.256328 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.383284 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt42d\" (UniqueName: \"kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.383774 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.384411 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.486174 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.486281 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt42d\" (UniqueName: \"kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.486505 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.486616 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.487209 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.506396 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt42d\" (UniqueName: \"kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d\") pod \"redhat-marketplace-24k4v\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:15 crc kubenswrapper[4688]: I1001 16:11:15.567026 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:16 crc kubenswrapper[4688]: W1001 16:11:16.038316 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98c19866_18b8_4bc4_9396_2dc96c249417.slice/crio-7ef617af54e03576ab197ef62770091be25c091d36c1c95080e90d9e39dfd03f WatchSource:0}: Error finding container 7ef617af54e03576ab197ef62770091be25c091d36c1c95080e90d9e39dfd03f: Status 404 returned error can't find the container with id 7ef617af54e03576ab197ef62770091be25c091d36c1c95080e90d9e39dfd03f Oct 01 16:11:16 crc kubenswrapper[4688]: I1001 16:11:16.040398 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:16 crc kubenswrapper[4688]: I1001 16:11:16.530904 4688 generic.go:334] "Generic (PLEG): container finished" podID="98c19866-18b8-4bc4-9396-2dc96c249417" containerID="709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88" exitCode=0 Oct 01 16:11:16 crc kubenswrapper[4688]: I1001 16:11:16.530943 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerDied","Data":"709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88"} Oct 01 16:11:16 crc kubenswrapper[4688]: I1001 16:11:16.532099 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerStarted","Data":"7ef617af54e03576ab197ef62770091be25c091d36c1c95080e90d9e39dfd03f"} Oct 01 16:11:18 crc kubenswrapper[4688]: I1001 16:11:18.555089 4688 generic.go:334] "Generic (PLEG): container finished" podID="98c19866-18b8-4bc4-9396-2dc96c249417" containerID="b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0" exitCode=0 Oct 01 16:11:18 crc kubenswrapper[4688]: I1001 16:11:18.555581 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerDied","Data":"b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0"} Oct 01 16:11:19 crc kubenswrapper[4688]: I1001 16:11:19.575539 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerStarted","Data":"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598"} Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.053114 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.053813 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.567244 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.567308 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.635774 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.660933 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-24k4v" podStartSLOduration=8.007637651 podStartE2EDuration="10.660912945s" podCreationTimestamp="2025-10-01 16:11:15 +0000 UTC" firstStartedPulling="2025-10-01 16:11:16.532666857 +0000 UTC m=+1465.883306829" lastFinishedPulling="2025-10-01 16:11:19.185942161 +0000 UTC m=+1468.536582123" observedRunningTime="2025-10-01 16:11:19.598016222 +0000 UTC m=+1468.948656224" watchObservedRunningTime="2025-10-01 16:11:25.660912945 +0000 UTC m=+1475.011552907" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.695773 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:25 crc kubenswrapper[4688]: I1001 16:11:25.876704 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:27 crc kubenswrapper[4688]: I1001 16:11:27.659724 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-24k4v" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="registry-server" containerID="cri-o://f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598" gracePeriod=2 Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.115840 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.244394 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities\") pod \"98c19866-18b8-4bc4-9396-2dc96c249417\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.245107 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities" (OuterVolumeSpecName: "utilities") pod "98c19866-18b8-4bc4-9396-2dc96c249417" (UID: "98c19866-18b8-4bc4-9396-2dc96c249417"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.245120 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt42d\" (UniqueName: \"kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d\") pod \"98c19866-18b8-4bc4-9396-2dc96c249417\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.245214 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content\") pod \"98c19866-18b8-4bc4-9396-2dc96c249417\" (UID: \"98c19866-18b8-4bc4-9396-2dc96c249417\") " Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.246113 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.256768 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d" (OuterVolumeSpecName: "kube-api-access-gt42d") pod "98c19866-18b8-4bc4-9396-2dc96c249417" (UID: "98c19866-18b8-4bc4-9396-2dc96c249417"). InnerVolumeSpecName "kube-api-access-gt42d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.258707 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98c19866-18b8-4bc4-9396-2dc96c249417" (UID: "98c19866-18b8-4bc4-9396-2dc96c249417"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.347817 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt42d\" (UniqueName: \"kubernetes.io/projected/98c19866-18b8-4bc4-9396-2dc96c249417-kube-api-access-gt42d\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.348001 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c19866-18b8-4bc4-9396-2dc96c249417-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.670726 4688 generic.go:334] "Generic (PLEG): container finished" podID="98c19866-18b8-4bc4-9396-2dc96c249417" containerID="f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598" exitCode=0 Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.670772 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerDied","Data":"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598"} Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.670812 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-24k4v" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.670838 4688 scope.go:117] "RemoveContainer" containerID="f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.670821 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-24k4v" event={"ID":"98c19866-18b8-4bc4-9396-2dc96c249417","Type":"ContainerDied","Data":"7ef617af54e03576ab197ef62770091be25c091d36c1c95080e90d9e39dfd03f"} Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.712261 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.714479 4688 scope.go:117] "RemoveContainer" containerID="b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.722443 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-24k4v"] Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.745205 4688 scope.go:117] "RemoveContainer" containerID="709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.800729 4688 scope.go:117] "RemoveContainer" containerID="f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598" Oct 01 16:11:28 crc kubenswrapper[4688]: E1001 16:11:28.801078 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598\": container with ID starting with f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598 not found: ID does not exist" containerID="f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.801161 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598"} err="failed to get container status \"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598\": rpc error: code = NotFound desc = could not find container \"f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598\": container with ID starting with f8c21f254e0e75c44c6884ed2bf7d8986604aa8e7736dcd52a08cafff7ba4598 not found: ID does not exist" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.801229 4688 scope.go:117] "RemoveContainer" containerID="b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0" Oct 01 16:11:28 crc kubenswrapper[4688]: E1001 16:11:28.801496 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0\": container with ID starting with b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0 not found: ID does not exist" containerID="b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.801586 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0"} err="failed to get container status \"b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0\": rpc error: code = NotFound desc = could not find container \"b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0\": container with ID starting with b4ae7c81a7bc13437e2d1c775f0f448413fbb268f124e8ff8190f73af18061e0 not found: ID does not exist" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.801655 4688 scope.go:117] "RemoveContainer" containerID="709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88" Oct 01 16:11:28 crc kubenswrapper[4688]: E1001 16:11:28.801876 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88\": container with ID starting with 709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88 not found: ID does not exist" containerID="709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88" Oct 01 16:11:28 crc kubenswrapper[4688]: I1001 16:11:28.801963 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88"} err="failed to get container status \"709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88\": rpc error: code = NotFound desc = could not find container \"709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88\": container with ID starting with 709a4bd29ee99d5c6968df6b51b73e264d732b770818c4aca9d8a7d33a738d88 not found: ID does not exist" Oct 01 16:11:29 crc kubenswrapper[4688]: I1001 16:11:29.395733 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" path="/var/lib/kubelet/pods/98c19866-18b8-4bc4-9396-2dc96c249417/volumes" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.645076 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-77ftd"] Oct 01 16:11:47 crc kubenswrapper[4688]: E1001 16:11:47.646112 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="extract-utilities" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.646129 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="extract-utilities" Oct 01 16:11:47 crc kubenswrapper[4688]: E1001 16:11:47.646155 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="extract-content" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.646164 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="extract-content" Oct 01 16:11:47 crc kubenswrapper[4688]: E1001 16:11:47.646184 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="registry-server" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.646192 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="registry-server" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.646460 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c19866-18b8-4bc4-9396-2dc96c249417" containerName="registry-server" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.648630 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.658994 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-77ftd"] Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.830011 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvzng\" (UniqueName: \"kubernetes.io/projected/25ac9f9e-1a94-4d95-bb20-20de73ca538a-kube-api-access-vvzng\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.830120 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-catalog-content\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.830218 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-utilities\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.931383 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-catalog-content\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.931788 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-utilities\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.931926 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-catalog-content\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.932019 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvzng\" (UniqueName: \"kubernetes.io/projected/25ac9f9e-1a94-4d95-bb20-20de73ca538a-kube-api-access-vvzng\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.932189 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ac9f9e-1a94-4d95-bb20-20de73ca538a-utilities\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.951045 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvzng\" (UniqueName: \"kubernetes.io/projected/25ac9f9e-1a94-4d95-bb20-20de73ca538a-kube-api-access-vvzng\") pod \"certified-operators-77ftd\" (UID: \"25ac9f9e-1a94-4d95-bb20-20de73ca538a\") " pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:47 crc kubenswrapper[4688]: I1001 16:11:47.983029 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:48 crc kubenswrapper[4688]: I1001 16:11:48.466913 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-77ftd"] Oct 01 16:11:48 crc kubenswrapper[4688]: I1001 16:11:48.868721 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77ftd" event={"ID":"25ac9f9e-1a94-4d95-bb20-20de73ca538a","Type":"ContainerStarted","Data":"e6ed19d7a8ee22cbb40621175d30fd64c1cfa58192dedec22df67d61227d7a19"} Oct 01 16:11:48 crc kubenswrapper[4688]: I1001 16:11:48.869153 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77ftd" event={"ID":"25ac9f9e-1a94-4d95-bb20-20de73ca538a","Type":"ContainerStarted","Data":"67164f64925725c6f2dac1a2719d4e12d425b3dba431190351c6b15aa14309f9"} Oct 01 16:11:49 crc kubenswrapper[4688]: I1001 16:11:49.879036 4688 generic.go:334] "Generic (PLEG): container finished" podID="25ac9f9e-1a94-4d95-bb20-20de73ca538a" containerID="e6ed19d7a8ee22cbb40621175d30fd64c1cfa58192dedec22df67d61227d7a19" exitCode=0 Oct 01 16:11:49 crc kubenswrapper[4688]: I1001 16:11:49.879143 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77ftd" event={"ID":"25ac9f9e-1a94-4d95-bb20-20de73ca538a","Type":"ContainerDied","Data":"e6ed19d7a8ee22cbb40621175d30fd64c1cfa58192dedec22df67d61227d7a19"} Oct 01 16:11:54 crc kubenswrapper[4688]: I1001 16:11:54.934906 4688 generic.go:334] "Generic (PLEG): container finished" podID="25ac9f9e-1a94-4d95-bb20-20de73ca538a" containerID="678d769c78a16080b42cb137843c23b5588289b6e8be3fc7ffa86af6ef60b951" exitCode=0 Oct 01 16:11:54 crc kubenswrapper[4688]: I1001 16:11:54.934982 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77ftd" event={"ID":"25ac9f9e-1a94-4d95-bb20-20de73ca538a","Type":"ContainerDied","Data":"678d769c78a16080b42cb137843c23b5588289b6e8be3fc7ffa86af6ef60b951"} Oct 01 16:11:55 crc kubenswrapper[4688]: I1001 16:11:55.052855 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:11:55 crc kubenswrapper[4688]: I1001 16:11:55.052942 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:11:55 crc kubenswrapper[4688]: I1001 16:11:55.053005 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:11:55 crc kubenswrapper[4688]: I1001 16:11:55.054026 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:11:55 crc kubenswrapper[4688]: I1001 16:11:55.054310 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" gracePeriod=600 Oct 01 16:11:56 crc kubenswrapper[4688]: E1001 16:11:56.339375 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:11:56 crc kubenswrapper[4688]: I1001 16:11:56.956119 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" exitCode=0 Oct 01 16:11:56 crc kubenswrapper[4688]: I1001 16:11:56.956181 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7"} Oct 01 16:11:56 crc kubenswrapper[4688]: I1001 16:11:56.956319 4688 scope.go:117] "RemoveContainer" containerID="b5262e411aa493ba503d6fd3de7edddc2425ba7e6bce100a9615e560ba93887c" Oct 01 16:11:56 crc kubenswrapper[4688]: I1001 16:11:56.956913 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:11:56 crc kubenswrapper[4688]: E1001 16:11:56.957314 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:11:56 crc kubenswrapper[4688]: I1001 16:11:56.959246 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-77ftd" event={"ID":"25ac9f9e-1a94-4d95-bb20-20de73ca538a","Type":"ContainerStarted","Data":"c9f0763ba009bff8ddc0ba5c92393f8424a4a18d0764c977dbbfb1db14996369"} Oct 01 16:11:57 crc kubenswrapper[4688]: I1001 16:11:57.005870 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-77ftd" podStartSLOduration=3.560266953 podStartE2EDuration="10.005851624s" podCreationTimestamp="2025-10-01 16:11:47 +0000 UTC" firstStartedPulling="2025-10-01 16:11:49.881051644 +0000 UTC m=+1499.231691616" lastFinishedPulling="2025-10-01 16:11:56.326636315 +0000 UTC m=+1505.677276287" observedRunningTime="2025-10-01 16:11:57.003322002 +0000 UTC m=+1506.353961984" watchObservedRunningTime="2025-10-01 16:11:57.005851624 +0000 UTC m=+1506.356491586" Oct 01 16:11:57 crc kubenswrapper[4688]: I1001 16:11:57.983328 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:57 crc kubenswrapper[4688]: I1001 16:11:57.983438 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:11:59 crc kubenswrapper[4688]: I1001 16:11:59.035133 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-77ftd" podUID="25ac9f9e-1a94-4d95-bb20-20de73ca538a" containerName="registry-server" probeResult="failure" output=< Oct 01 16:11:59 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:11:59 crc kubenswrapper[4688]: > Oct 01 16:11:59 crc kubenswrapper[4688]: I1001 16:11:59.216428 4688 scope.go:117] "RemoveContainer" containerID="d4a9a43158e938fbb69326381e5b182bdb47e0f1e8f611b017e1bee9b41d1c0c" Oct 01 16:12:08 crc kubenswrapper[4688]: I1001 16:12:08.045469 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:12:08 crc kubenswrapper[4688]: I1001 16:12:08.104951 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-77ftd" Oct 01 16:12:08 crc kubenswrapper[4688]: I1001 16:12:08.192576 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-77ftd"] Oct 01 16:12:08 crc kubenswrapper[4688]: I1001 16:12:08.333605 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 16:12:08 crc kubenswrapper[4688]: I1001 16:12:08.333961 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w5tln" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="registry-server" containerID="cri-o://9baddb6f9d1506db1adbec28541f0d94ea4a8d2e5d1e7b284a0f804dc7d26402" gracePeriod=2 Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.088907 4688 generic.go:334] "Generic (PLEG): container finished" podID="245214e6-aa79-4963-9524-2c6852196d70" containerID="9baddb6f9d1506db1adbec28541f0d94ea4a8d2e5d1e7b284a0f804dc7d26402" exitCode=0 Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.088971 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerDied","Data":"9baddb6f9d1506db1adbec28541f0d94ea4a8d2e5d1e7b284a0f804dc7d26402"} Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.428346 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.484365 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ngmc\" (UniqueName: \"kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc\") pod \"245214e6-aa79-4963-9524-2c6852196d70\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.484540 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content\") pod \"245214e6-aa79-4963-9524-2c6852196d70\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.484651 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities\") pod \"245214e6-aa79-4963-9524-2c6852196d70\" (UID: \"245214e6-aa79-4963-9524-2c6852196d70\") " Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.490083 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities" (OuterVolumeSpecName: "utilities") pod "245214e6-aa79-4963-9524-2c6852196d70" (UID: "245214e6-aa79-4963-9524-2c6852196d70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.495318 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc" (OuterVolumeSpecName: "kube-api-access-7ngmc") pod "245214e6-aa79-4963-9524-2c6852196d70" (UID: "245214e6-aa79-4963-9524-2c6852196d70"). InnerVolumeSpecName "kube-api-access-7ngmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.585508 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "245214e6-aa79-4963-9524-2c6852196d70" (UID: "245214e6-aa79-4963-9524-2c6852196d70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.586948 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ngmc\" (UniqueName: \"kubernetes.io/projected/245214e6-aa79-4963-9524-2c6852196d70-kube-api-access-7ngmc\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.586997 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:09 crc kubenswrapper[4688]: I1001 16:12:09.587011 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245214e6-aa79-4963-9524-2c6852196d70-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.103372 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5tln" event={"ID":"245214e6-aa79-4963-9524-2c6852196d70","Type":"ContainerDied","Data":"ad3d10906e73b8eec77cec519812024817370fc9bcd9a2efb5638289b8c1ac96"} Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.103594 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5tln" Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.103658 4688 scope.go:117] "RemoveContainer" containerID="9baddb6f9d1506db1adbec28541f0d94ea4a8d2e5d1e7b284a0f804dc7d26402" Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.160753 4688 scope.go:117] "RemoveContainer" containerID="59363691d0fbd2163afb8331702d6ec613069b8716ffc4ea38af62a20a7a11ff" Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.162274 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.185306 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w5tln"] Oct 01 16:12:10 crc kubenswrapper[4688]: I1001 16:12:10.218828 4688 scope.go:117] "RemoveContainer" containerID="aa5178bc6063b3df9716bd3f7f2317daaf045c2a33f5d6c49c3b2d8988105f5e" Oct 01 16:12:11 crc kubenswrapper[4688]: I1001 16:12:11.388858 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:12:11 crc kubenswrapper[4688]: E1001 16:12:11.389434 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:12:11 crc kubenswrapper[4688]: I1001 16:12:11.394021 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245214e6-aa79-4963-9524-2c6852196d70" path="/var/lib/kubelet/pods/245214e6-aa79-4963-9524-2c6852196d70/volumes" Oct 01 16:12:24 crc kubenswrapper[4688]: I1001 16:12:24.381700 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:12:24 crc kubenswrapper[4688]: E1001 16:12:24.382665 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:12:35 crc kubenswrapper[4688]: I1001 16:12:35.381605 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:12:35 crc kubenswrapper[4688]: E1001 16:12:35.382573 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:12:46 crc kubenswrapper[4688]: I1001 16:12:46.381817 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:12:46 crc kubenswrapper[4688]: E1001 16:12:46.382761 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:12:53 crc kubenswrapper[4688]: I1001 16:12:53.364217 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-6bbc5fdc47-8258p" podUID="43929b9b-990a-4a85-9100-509d80bba3c1" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 16:12:58 crc kubenswrapper[4688]: I1001 16:12:58.381292 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:12:58 crc kubenswrapper[4688]: E1001 16:12:58.382052 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:12:59 crc kubenswrapper[4688]: I1001 16:12:59.307830 4688 scope.go:117] "RemoveContainer" containerID="f630d5f9a142198bb5b70976cb8d5c6e450a2cc7c8451370467ffd9e3ad2d12c" Oct 01 16:12:59 crc kubenswrapper[4688]: I1001 16:12:59.352622 4688 scope.go:117] "RemoveContainer" containerID="58f1884793ca8c394b4449d587709fe97e059bdbb69059104602956929edf013" Oct 01 16:12:59 crc kubenswrapper[4688]: I1001 16:12:59.393073 4688 scope.go:117] "RemoveContainer" containerID="f87044b84436f2a688c3f949cb91e14f043f3c4c85739aade57d998dac4b2ff4" Oct 01 16:13:07 crc kubenswrapper[4688]: I1001 16:13:07.061583 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7f469"] Oct 01 16:13:07 crc kubenswrapper[4688]: I1001 16:13:07.074641 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7f469"] Oct 01 16:13:07 crc kubenswrapper[4688]: I1001 16:13:07.398192 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88217dd8-e3f4-4a77-8960-80980cda5e6a" path="/var/lib/kubelet/pods/88217dd8-e3f4-4a77-8960-80980cda5e6a/volumes" Oct 01 16:13:09 crc kubenswrapper[4688]: I1001 16:13:09.380617 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:13:09 crc kubenswrapper[4688]: E1001 16:13:09.381003 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:13:13 crc kubenswrapper[4688]: I1001 16:13:13.043762 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lxw94"] Oct 01 16:13:13 crc kubenswrapper[4688]: I1001 16:13:13.055596 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lxw94"] Oct 01 16:13:13 crc kubenswrapper[4688]: I1001 16:13:13.464782 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7dff43a-da4b-412d-a08e-eae2053bfb8c" path="/var/lib/kubelet/pods/d7dff43a-da4b-412d-a08e-eae2053bfb8c/volumes" Oct 01 16:13:14 crc kubenswrapper[4688]: I1001 16:13:14.036161 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4grcg"] Oct 01 16:13:14 crc kubenswrapper[4688]: I1001 16:13:14.049178 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4grcg"] Oct 01 16:13:15 crc kubenswrapper[4688]: I1001 16:13:15.392434 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5227ef09-55af-48c6-a8ee-e6ff58bca51a" path="/var/lib/kubelet/pods/5227ef09-55af-48c6-a8ee-e6ff58bca51a/volumes" Oct 01 16:13:23 crc kubenswrapper[4688]: I1001 16:13:23.383480 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:13:23 crc kubenswrapper[4688]: E1001 16:13:23.384483 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.031323 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6bb3-account-create-8hlc5"] Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.042561 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8661-account-create-q74m4"] Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.049933 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-570c-account-create-m78fq"] Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.057158 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6bb3-account-create-8hlc5"] Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.065374 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8661-account-create-q74m4"] Oct 01 16:13:26 crc kubenswrapper[4688]: I1001 16:13:26.072155 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-570c-account-create-m78fq"] Oct 01 16:13:27 crc kubenswrapper[4688]: I1001 16:13:27.413394 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05a1749d-a3be-4a23-888f-7ace91e261c2" path="/var/lib/kubelet/pods/05a1749d-a3be-4a23-888f-7ace91e261c2/volumes" Oct 01 16:13:27 crc kubenswrapper[4688]: I1001 16:13:27.417182 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="568ac047-f5a3-4d4b-bf31-9a294e0ccf6a" path="/var/lib/kubelet/pods/568ac047-f5a3-4d4b-bf31-9a294e0ccf6a/volumes" Oct 01 16:13:27 crc kubenswrapper[4688]: I1001 16:13:27.426138 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c43636cd-b73f-4427-860a-7d9c40f0732a" path="/var/lib/kubelet/pods/c43636cd-b73f-4427-860a-7d9c40f0732a/volumes" Oct 01 16:13:34 crc kubenswrapper[4688]: I1001 16:13:34.381609 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:13:34 crc kubenswrapper[4688]: E1001 16:13:34.382338 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:13:46 crc kubenswrapper[4688]: I1001 16:13:46.380945 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:13:46 crc kubenswrapper[4688]: E1001 16:13:46.381955 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.045540 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-nhpdd"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.057657 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-dq424"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.069316 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bqz97"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.078643 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-nhpdd"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.086719 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bqz97"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.099983 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-dq424"] Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.141673 4688 generic.go:334] "Generic (PLEG): container finished" podID="fbd7ce66-5861-4a86-a892-ce6493baf848" containerID="485b7601e0bf7f5860dca83c9434747bfffa78a596ffbf41abb01500e1affbfc" exitCode=0 Oct 01 16:13:52 crc kubenswrapper[4688]: I1001 16:13:52.141730 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" event={"ID":"fbd7ce66-5861-4a86-a892-ce6493baf848","Type":"ContainerDied","Data":"485b7601e0bf7f5860dca83c9434747bfffa78a596ffbf41abb01500e1affbfc"} Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.422206 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e" path="/var/lib/kubelet/pods/7e9d755e-6f97-4ec9-ac9c-1ac17de86d6e/volumes" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.425160 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb2febea-70d3-47ed-9ee2-1daf35d13f86" path="/var/lib/kubelet/pods/bb2febea-70d3-47ed-9ee2-1daf35d13f86/volumes" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.427808 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdacba2-e02a-4105-bd69-8791577b722c" path="/var/lib/kubelet/pods/efdacba2-e02a-4105-bd69-8791577b722c/volumes" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.547908 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.744851 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory\") pod \"fbd7ce66-5861-4a86-a892-ce6493baf848\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.745331 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key\") pod \"fbd7ce66-5861-4a86-a892-ce6493baf848\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.745419 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle\") pod \"fbd7ce66-5861-4a86-a892-ce6493baf848\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.745570 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfms5\" (UniqueName: \"kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5\") pod \"fbd7ce66-5861-4a86-a892-ce6493baf848\" (UID: \"fbd7ce66-5861-4a86-a892-ce6493baf848\") " Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.750421 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5" (OuterVolumeSpecName: "kube-api-access-bfms5") pod "fbd7ce66-5861-4a86-a892-ce6493baf848" (UID: "fbd7ce66-5861-4a86-a892-ce6493baf848"). InnerVolumeSpecName "kube-api-access-bfms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.750792 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fbd7ce66-5861-4a86-a892-ce6493baf848" (UID: "fbd7ce66-5861-4a86-a892-ce6493baf848"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.775058 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory" (OuterVolumeSpecName: "inventory") pod "fbd7ce66-5861-4a86-a892-ce6493baf848" (UID: "fbd7ce66-5861-4a86-a892-ce6493baf848"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.779063 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fbd7ce66-5861-4a86-a892-ce6493baf848" (UID: "fbd7ce66-5861-4a86-a892-ce6493baf848"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.847621 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfms5\" (UniqueName: \"kubernetes.io/projected/fbd7ce66-5861-4a86-a892-ce6493baf848-kube-api-access-bfms5\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.847659 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.847672 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:53 crc kubenswrapper[4688]: I1001 16:13:53.847684 4688 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7ce66-5861-4a86-a892-ce6493baf848-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.160014 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" event={"ID":"fbd7ce66-5861-4a86-a892-ce6493baf848","Type":"ContainerDied","Data":"caf3344f3173d26242eb545f7ec03cc0d582ecf589cd1d6bae6fc6aa91b09457"} Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.160057 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caf3344f3173d26242eb545f7ec03cc0d582ecf589cd1d6bae6fc6aa91b09457" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.160110 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bk825" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.266293 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f"] Oct 01 16:13:54 crc kubenswrapper[4688]: E1001 16:13:54.266807 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="extract-utilities" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.266829 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="extract-utilities" Oct 01 16:13:54 crc kubenswrapper[4688]: E1001 16:13:54.266860 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="extract-content" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.266868 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="extract-content" Oct 01 16:13:54 crc kubenswrapper[4688]: E1001 16:13:54.266878 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="registry-server" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.266887 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="registry-server" Oct 01 16:13:54 crc kubenswrapper[4688]: E1001 16:13:54.266919 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd7ce66-5861-4a86-a892-ce6493baf848" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.266928 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd7ce66-5861-4a86-a892-ce6493baf848" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.267143 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="245214e6-aa79-4963-9524-2c6852196d70" containerName="registry-server" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.267165 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd7ce66-5861-4a86-a892-ce6493baf848" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.267931 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.272789 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.272818 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.272800 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.272902 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.286434 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f"] Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.457814 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.457898 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8nm\" (UniqueName: \"kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.457938 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.559515 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.559588 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8nm\" (UniqueName: \"kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.559613 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.564666 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.572662 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.580425 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8nm\" (UniqueName: \"kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:54 crc kubenswrapper[4688]: I1001 16:13:54.585401 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:13:55 crc kubenswrapper[4688]: I1001 16:13:55.124965 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f"] Oct 01 16:13:55 crc kubenswrapper[4688]: I1001 16:13:55.129781 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:13:55 crc kubenswrapper[4688]: I1001 16:13:55.170350 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" event={"ID":"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe","Type":"ContainerStarted","Data":"0bb1b85b42a93915c2a4dfb20342d8b8ce75a4012bbffc634b27ba1b47d036e3"} Oct 01 16:13:56 crc kubenswrapper[4688]: I1001 16:13:56.177881 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" event={"ID":"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe","Type":"ContainerStarted","Data":"b54c3be600e557fc5a143f63318ad4f0fd8e199d5d46cb25c31f7e0ba41d6317"} Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.510240 4688 scope.go:117] "RemoveContainer" containerID="4246ab03e0c4431300f6451df412d3c177eef48fd8b985b6b34fef0d3019c9ec" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.538383 4688 scope.go:117] "RemoveContainer" containerID="1765e397f8877587e83a24018aafd9a6d3fb2ef46b0dd4c821066f969842871a" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.596829 4688 scope.go:117] "RemoveContainer" containerID="80ec9296ac066bb97c864f35192c472ed3c0638029261e21b156f659d48a6c3f" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.650559 4688 scope.go:117] "RemoveContainer" containerID="67e47d68212350a1d0a70a8831ae241af4c59011b5c34123b1e0c6d0d5619793" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.691749 4688 scope.go:117] "RemoveContainer" containerID="47344ebf8c4c49c75b30b173377e7ba797c4caf38bd15fade03c857f475ee394" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.753667 4688 scope.go:117] "RemoveContainer" containerID="1fd84cc04e29376ad465e8deea794a6aa5f8df429a4da30283ffba5038ec8ec5" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.807243 4688 scope.go:117] "RemoveContainer" containerID="0e92bafaef8a011fbf78e883dce26a5cd1669424d6ca178c3ba37e6bd02e1bb9" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.835356 4688 scope.go:117] "RemoveContainer" containerID="669b5751d549abf566560b9bb60f58c3f0b197455f9e49bded7b00a72df35f21" Oct 01 16:13:59 crc kubenswrapper[4688]: I1001 16:13:59.856968 4688 scope.go:117] "RemoveContainer" containerID="a1b0b43b9fdaf60a9ee4a24a49a01d3ba5423fea699d88c2dbf7c00bef2b06f7" Oct 01 16:14:01 crc kubenswrapper[4688]: I1001 16:14:01.032380 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" podStartSLOduration=6.538883393 podStartE2EDuration="7.032356886s" podCreationTimestamp="2025-10-01 16:13:54 +0000 UTC" firstStartedPulling="2025-10-01 16:13:55.129439425 +0000 UTC m=+1624.480079387" lastFinishedPulling="2025-10-01 16:13:55.622912918 +0000 UTC m=+1624.973552880" observedRunningTime="2025-10-01 16:13:56.197333024 +0000 UTC m=+1625.547972986" watchObservedRunningTime="2025-10-01 16:14:01.032356886 +0000 UTC m=+1630.382996858" Oct 01 16:14:01 crc kubenswrapper[4688]: I1001 16:14:01.035068 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-52dc-account-create-jqvjq"] Oct 01 16:14:01 crc kubenswrapper[4688]: I1001 16:14:01.044869 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-52dc-account-create-jqvjq"] Oct 01 16:14:01 crc kubenswrapper[4688]: I1001 16:14:01.387377 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:14:01 crc kubenswrapper[4688]: E1001 16:14:01.387678 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:14:01 crc kubenswrapper[4688]: I1001 16:14:01.396892 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59228674-5e5c-426f-9e2d-e82c5760b356" path="/var/lib/kubelet/pods/59228674-5e5c-426f-9e2d-e82c5760b356/volumes" Oct 01 16:14:02 crc kubenswrapper[4688]: I1001 16:14:02.031113 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d7e0-account-create-58kmq"] Oct 01 16:14:02 crc kubenswrapper[4688]: I1001 16:14:02.037967 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d7e0-account-create-58kmq"] Oct 01 16:14:03 crc kubenswrapper[4688]: I1001 16:14:03.043109 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-058d-account-create-fzsds"] Oct 01 16:14:03 crc kubenswrapper[4688]: I1001 16:14:03.052314 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-058d-account-create-fzsds"] Oct 01 16:14:03 crc kubenswrapper[4688]: I1001 16:14:03.398226 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce461b22-96d9-42ee-950d-b001054c5c60" path="/var/lib/kubelet/pods/ce461b22-96d9-42ee-950d-b001054c5c60/volumes" Oct 01 16:14:03 crc kubenswrapper[4688]: I1001 16:14:03.399631 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c4d47a-f834-4952-a414-34fb8df2a1ef" path="/var/lib/kubelet/pods/f8c4d47a-f834-4952-a414-34fb8df2a1ef/volumes" Oct 01 16:14:04 crc kubenswrapper[4688]: I1001 16:14:04.053366 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bk86d"] Oct 01 16:14:04 crc kubenswrapper[4688]: I1001 16:14:04.070003 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bk86d"] Oct 01 16:14:05 crc kubenswrapper[4688]: I1001 16:14:05.051133 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-l5md4"] Oct 01 16:14:05 crc kubenswrapper[4688]: I1001 16:14:05.067183 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-l5md4"] Oct 01 16:14:05 crc kubenswrapper[4688]: I1001 16:14:05.390623 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad6973b9-8d39-441f-ae39-67b632e06bcf" path="/var/lib/kubelet/pods/ad6973b9-8d39-441f-ae39-67b632e06bcf/volumes" Oct 01 16:14:05 crc kubenswrapper[4688]: I1001 16:14:05.392135 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e322d157-d4b5-4b2c-92a6-141481c2b358" path="/var/lib/kubelet/pods/e322d157-d4b5-4b2c-92a6-141481c2b358/volumes" Oct 01 16:14:14 crc kubenswrapper[4688]: I1001 16:14:14.382369 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:14:14 crc kubenswrapper[4688]: E1001 16:14:14.383118 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:14:28 crc kubenswrapper[4688]: I1001 16:14:28.381293 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:14:28 crc kubenswrapper[4688]: E1001 16:14:28.382138 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:14:41 crc kubenswrapper[4688]: I1001 16:14:41.386882 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:14:41 crc kubenswrapper[4688]: E1001 16:14:41.387902 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:14:44 crc kubenswrapper[4688]: I1001 16:14:44.058314 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-flx64"] Oct 01 16:14:44 crc kubenswrapper[4688]: I1001 16:14:44.066761 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-flx64"] Oct 01 16:14:45 crc kubenswrapper[4688]: I1001 16:14:45.401841 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a989728-c717-4967-a372-93152bd2f064" path="/var/lib/kubelet/pods/4a989728-c717-4967-a372-93152bd2f064/volumes" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.191267 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.194193 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.212546 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.379710 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.381254 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6v5k\" (UniqueName: \"kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.381355 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.483009 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.483128 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6v5k\" (UniqueName: \"kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.483156 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.483873 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.484298 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.508497 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6v5k\" (UniqueName: \"kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k\") pod \"redhat-operators-7t2zl\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:46 crc kubenswrapper[4688]: I1001 16:14:46.523650 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.018593 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.665193 4688 generic.go:334] "Generic (PLEG): container finished" podID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerID="fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958" exitCode=0 Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.665292 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerDied","Data":"fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958"} Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.666869 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerStarted","Data":"69b414e51549264efa64e6385b497486e852742480a08c8bfad7b9cf12968027"} Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.794012 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.796732 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.803492 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.910137 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.910186 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:47 crc kubenswrapper[4688]: I1001 16:14:47.910231 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbltl\" (UniqueName: \"kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.012494 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.012572 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.012618 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbltl\" (UniqueName: \"kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.013108 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.013256 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.038360 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbltl\" (UniqueName: \"kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl\") pod \"community-operators-6lgp4\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.154236 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:48 crc kubenswrapper[4688]: I1001 16:14:48.678858 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:14:49 crc kubenswrapper[4688]: I1001 16:14:49.695876 4688 generic.go:334] "Generic (PLEG): container finished" podID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerID="a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657" exitCode=0 Oct 01 16:14:49 crc kubenswrapper[4688]: I1001 16:14:49.695938 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerDied","Data":"a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657"} Oct 01 16:14:49 crc kubenswrapper[4688]: I1001 16:14:49.696377 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerStarted","Data":"24823cbb6ffdced608281d3b452631ea5f525ab4a5e5f2c4ce927b7599b5b19e"} Oct 01 16:14:49 crc kubenswrapper[4688]: I1001 16:14:49.703575 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerStarted","Data":"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3"} Oct 01 16:14:51 crc kubenswrapper[4688]: I1001 16:14:51.718654 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerStarted","Data":"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b"} Oct 01 16:14:54 crc kubenswrapper[4688]: I1001 16:14:54.028487 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ntpwg"] Oct 01 16:14:54 crc kubenswrapper[4688]: I1001 16:14:54.043905 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ntpwg"] Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.030488 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-82bf8"] Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.046010 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-82bf8"] Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.394706 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e68b63-d053-412d-95e8-0e5f10bffc45" path="/var/lib/kubelet/pods/26e68b63-d053-412d-95e8-0e5f10bffc45/volumes" Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.396993 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94f981e-54a9-4bca-ad26-518b47ca0fb3" path="/var/lib/kubelet/pods/f94f981e-54a9-4bca-ad26-518b47ca0fb3/volumes" Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.751356 4688 generic.go:334] "Generic (PLEG): container finished" podID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerID="fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b" exitCode=0 Oct 01 16:14:55 crc kubenswrapper[4688]: I1001 16:14:55.751444 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerDied","Data":"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b"} Oct 01 16:14:56 crc kubenswrapper[4688]: I1001 16:14:56.381121 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:14:56 crc kubenswrapper[4688]: E1001 16:14:56.381454 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:14:57 crc kubenswrapper[4688]: I1001 16:14:57.771958 4688 generic.go:334] "Generic (PLEG): container finished" podID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerID="ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3" exitCode=0 Oct 01 16:14:57 crc kubenswrapper[4688]: I1001 16:14:57.772017 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerDied","Data":"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3"} Oct 01 16:14:57 crc kubenswrapper[4688]: I1001 16:14:57.774484 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerStarted","Data":"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757"} Oct 01 16:14:57 crc kubenswrapper[4688]: I1001 16:14:57.822039 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6lgp4" podStartSLOduration=3.782558749 podStartE2EDuration="10.822018299s" podCreationTimestamp="2025-10-01 16:14:47 +0000 UTC" firstStartedPulling="2025-10-01 16:14:49.703724938 +0000 UTC m=+1679.054364900" lastFinishedPulling="2025-10-01 16:14:56.743184488 +0000 UTC m=+1686.093824450" observedRunningTime="2025-10-01 16:14:57.819125267 +0000 UTC m=+1687.169765249" watchObservedRunningTime="2025-10-01 16:14:57.822018299 +0000 UTC m=+1687.172658261" Oct 01 16:14:58 crc kubenswrapper[4688]: I1001 16:14:58.155349 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:58 crc kubenswrapper[4688]: I1001 16:14:58.155415 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:14:58 crc kubenswrapper[4688]: I1001 16:14:58.784727 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerStarted","Data":"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586"} Oct 01 16:14:58 crc kubenswrapper[4688]: I1001 16:14:58.808463 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7t2zl" podStartSLOduration=2.110686403 podStartE2EDuration="12.808441348s" podCreationTimestamp="2025-10-01 16:14:46 +0000 UTC" firstStartedPulling="2025-10-01 16:14:47.668028488 +0000 UTC m=+1677.018668450" lastFinishedPulling="2025-10-01 16:14:58.365783423 +0000 UTC m=+1687.716423395" observedRunningTime="2025-10-01 16:14:58.806651457 +0000 UTC m=+1688.157291439" watchObservedRunningTime="2025-10-01 16:14:58.808441348 +0000 UTC m=+1688.159081310" Oct 01 16:14:59 crc kubenswrapper[4688]: I1001 16:14:59.200424 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6lgp4" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="registry-server" probeResult="failure" output=< Oct 01 16:14:59 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:14:59 crc kubenswrapper[4688]: > Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.091030 4688 scope.go:117] "RemoveContainer" containerID="1a7f60f52a6e0cf131857189106dac12916aaba97dd111722dbbb63576616a99" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.117275 4688 scope.go:117] "RemoveContainer" containerID="8c779883d99658a75782ccb222ae4e75028da3290c76740c0708a162626cab76" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.155533 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r"] Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.157476 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.160670 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.161959 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.169392 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r"] Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.203290 4688 scope.go:117] "RemoveContainer" containerID="de3e989d53a2da480bed2e6dae2272057e53c483fce3ba61e00724c73afff06b" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.252727 4688 scope.go:117] "RemoveContainer" containerID="0d0979eba270753cd514b797174ad6d89b0308dcdacd47698c0539f0061871cb" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.336414 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.336596 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.336673 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrdzg\" (UniqueName: \"kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.355957 4688 scope.go:117] "RemoveContainer" containerID="4c1182b103b63c55e821d4c81e0bc842973f8d71af838535e1aa1cd95606d7f5" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.393643 4688 scope.go:117] "RemoveContainer" containerID="b5de4ddd09f876e9803f292b21becf10b4aa56aab7370c7d20878f1a40ecc847" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.438730 4688 scope.go:117] "RemoveContainer" containerID="e280f20963ee60ed0f9b24bb156990c57adfec45b97ee9c49d49a285630e73fa" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.438808 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.438870 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.438923 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrdzg\" (UniqueName: \"kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.439944 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.452903 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.462384 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrdzg\" (UniqueName: \"kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg\") pod \"collect-profiles-29322255-4kv6r\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.462620 4688 scope.go:117] "RemoveContainer" containerID="46fcbc2ddeab50f23f49d9c431eb49e6459975bf3203d72e65912f21aa081c8a" Oct 01 16:15:00 crc kubenswrapper[4688]: I1001 16:15:00.572468 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:01 crc kubenswrapper[4688]: I1001 16:15:01.092222 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r"] Oct 01 16:15:01 crc kubenswrapper[4688]: W1001 16:15:01.101675 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda3c1bd4_a0a9_4778_8797_45bf6d6d017d.slice/crio-b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d WatchSource:0}: Error finding container b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d: Status 404 returned error can't find the container with id b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d Oct 01 16:15:01 crc kubenswrapper[4688]: I1001 16:15:01.816760 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" event={"ID":"da3c1bd4-a0a9-4778-8797-45bf6d6d017d","Type":"ContainerStarted","Data":"518ec85f7d564863dadfeddc5a8585cffee2d83df970530729ed5c4be0c82822"} Oct 01 16:15:01 crc kubenswrapper[4688]: I1001 16:15:01.817063 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" event={"ID":"da3c1bd4-a0a9-4778-8797-45bf6d6d017d","Type":"ContainerStarted","Data":"b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d"} Oct 01 16:15:01 crc kubenswrapper[4688]: I1001 16:15:01.836809 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" podStartSLOduration=1.836793095 podStartE2EDuration="1.836793095s" podCreationTimestamp="2025-10-01 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:15:01.832324869 +0000 UTC m=+1691.182964831" watchObservedRunningTime="2025-10-01 16:15:01.836793095 +0000 UTC m=+1691.187433057" Oct 01 16:15:02 crc kubenswrapper[4688]: I1001 16:15:02.840011 4688 generic.go:334] "Generic (PLEG): container finished" podID="da3c1bd4-a0a9-4778-8797-45bf6d6d017d" containerID="518ec85f7d564863dadfeddc5a8585cffee2d83df970530729ed5c4be0c82822" exitCode=0 Oct 01 16:15:02 crc kubenswrapper[4688]: I1001 16:15:02.840479 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" event={"ID":"da3c1bd4-a0a9-4778-8797-45bf6d6d017d","Type":"ContainerDied","Data":"518ec85f7d564863dadfeddc5a8585cffee2d83df970530729ed5c4be0c82822"} Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.206113 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.324094 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume\") pod \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.324311 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume\") pod \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.324352 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrdzg\" (UniqueName: \"kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg\") pod \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\" (UID: \"da3c1bd4-a0a9-4778-8797-45bf6d6d017d\") " Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.325179 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume" (OuterVolumeSpecName: "config-volume") pod "da3c1bd4-a0a9-4778-8797-45bf6d6d017d" (UID: "da3c1bd4-a0a9-4778-8797-45bf6d6d017d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.332955 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg" (OuterVolumeSpecName: "kube-api-access-vrdzg") pod "da3c1bd4-a0a9-4778-8797-45bf6d6d017d" (UID: "da3c1bd4-a0a9-4778-8797-45bf6d6d017d"). InnerVolumeSpecName "kube-api-access-vrdzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.333326 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "da3c1bd4-a0a9-4778-8797-45bf6d6d017d" (UID: "da3c1bd4-a0a9-4778-8797-45bf6d6d017d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.426875 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrdzg\" (UniqueName: \"kubernetes.io/projected/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-kube-api-access-vrdzg\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.427159 4688 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.427303 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da3c1bd4-a0a9-4778-8797-45bf6d6d017d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.860191 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" event={"ID":"da3c1bd4-a0a9-4778-8797-45bf6d6d017d","Type":"ContainerDied","Data":"b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d"} Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.860689 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b00a8fe7e050ee9b2e8e7be106f3ef1125f64f60d930f95a600f837afaf1357d" Oct 01 16:15:04 crc kubenswrapper[4688]: I1001 16:15:04.860427 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322255-4kv6r" Oct 01 16:15:06 crc kubenswrapper[4688]: I1001 16:15:06.523965 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:06 crc kubenswrapper[4688]: I1001 16:15:06.524252 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:07 crc kubenswrapper[4688]: I1001 16:15:07.025313 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-bpllf"] Oct 01 16:15:07 crc kubenswrapper[4688]: I1001 16:15:07.032909 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-bpllf"] Oct 01 16:15:07 crc kubenswrapper[4688]: I1001 16:15:07.381733 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:15:07 crc kubenswrapper[4688]: E1001 16:15:07.382458 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:15:07 crc kubenswrapper[4688]: I1001 16:15:07.393439 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ff202f4-494c-49d4-a228-2e85d6659bb9" path="/var/lib/kubelet/pods/8ff202f4-494c-49d4-a228-2e85d6659bb9/volumes" Oct 01 16:15:07 crc kubenswrapper[4688]: I1001 16:15:07.572994 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7t2zl" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="registry-server" probeResult="failure" output=< Oct 01 16:15:07 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:15:07 crc kubenswrapper[4688]: > Oct 01 16:15:08 crc kubenswrapper[4688]: I1001 16:15:08.032503 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-bqw9p"] Oct 01 16:15:08 crc kubenswrapper[4688]: I1001 16:15:08.044385 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-bqw9p"] Oct 01 16:15:08 crc kubenswrapper[4688]: I1001 16:15:08.205459 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:15:08 crc kubenswrapper[4688]: I1001 16:15:08.257255 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:15:08 crc kubenswrapper[4688]: I1001 16:15:08.443224 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:15:09 crc kubenswrapper[4688]: I1001 16:15:09.393073 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81b3c008-7340-436e-873b-aa0d41588c6c" path="/var/lib/kubelet/pods/81b3c008-7340-436e-873b-aa0d41588c6c/volumes" Oct 01 16:15:09 crc kubenswrapper[4688]: I1001 16:15:09.905095 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6lgp4" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="registry-server" containerID="cri-o://673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757" gracePeriod=2 Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.404456 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.561994 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content\") pod \"a7e0b219-1147-4212-9a01-2d66a24601d7\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.562994 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities\") pod \"a7e0b219-1147-4212-9a01-2d66a24601d7\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.563029 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities" (OuterVolumeSpecName: "utilities") pod "a7e0b219-1147-4212-9a01-2d66a24601d7" (UID: "a7e0b219-1147-4212-9a01-2d66a24601d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.563279 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbltl\" (UniqueName: \"kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl\") pod \"a7e0b219-1147-4212-9a01-2d66a24601d7\" (UID: \"a7e0b219-1147-4212-9a01-2d66a24601d7\") " Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.564637 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.575106 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl" (OuterVolumeSpecName: "kube-api-access-sbltl") pod "a7e0b219-1147-4212-9a01-2d66a24601d7" (UID: "a7e0b219-1147-4212-9a01-2d66a24601d7"). InnerVolumeSpecName "kube-api-access-sbltl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.614572 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7e0b219-1147-4212-9a01-2d66a24601d7" (UID: "a7e0b219-1147-4212-9a01-2d66a24601d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.666650 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbltl\" (UniqueName: \"kubernetes.io/projected/a7e0b219-1147-4212-9a01-2d66a24601d7-kube-api-access-sbltl\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.666691 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e0b219-1147-4212-9a01-2d66a24601d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.918880 4688 generic.go:334] "Generic (PLEG): container finished" podID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerID="673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757" exitCode=0 Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.918955 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6lgp4" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.918980 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerDied","Data":"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757"} Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.919308 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6lgp4" event={"ID":"a7e0b219-1147-4212-9a01-2d66a24601d7","Type":"ContainerDied","Data":"24823cbb6ffdced608281d3b452631ea5f525ab4a5e5f2c4ce927b7599b5b19e"} Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.919335 4688 scope.go:117] "RemoveContainer" containerID="673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.961234 4688 scope.go:117] "RemoveContainer" containerID="fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b" Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.963636 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.977938 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6lgp4"] Oct 01 16:15:10 crc kubenswrapper[4688]: I1001 16:15:10.989137 4688 scope.go:117] "RemoveContainer" containerID="a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.043025 4688 scope.go:117] "RemoveContainer" containerID="673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757" Oct 01 16:15:11 crc kubenswrapper[4688]: E1001 16:15:11.043810 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757\": container with ID starting with 673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757 not found: ID does not exist" containerID="673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.043939 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757"} err="failed to get container status \"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757\": rpc error: code = NotFound desc = could not find container \"673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757\": container with ID starting with 673ab97027eac4fca5b723485404ba35112950d548d391302e024ce257c10757 not found: ID does not exist" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.044044 4688 scope.go:117] "RemoveContainer" containerID="fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b" Oct 01 16:15:11 crc kubenswrapper[4688]: E1001 16:15:11.045936 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b\": container with ID starting with fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b not found: ID does not exist" containerID="fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.046113 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b"} err="failed to get container status \"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b\": rpc error: code = NotFound desc = could not find container \"fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b\": container with ID starting with fac2c4b5fb22cc9f49ffc974c572938b6294f247a5895fa0866bb6bf75f6173b not found: ID does not exist" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.046218 4688 scope.go:117] "RemoveContainer" containerID="a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657" Oct 01 16:15:11 crc kubenswrapper[4688]: E1001 16:15:11.046693 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657\": container with ID starting with a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657 not found: ID does not exist" containerID="a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.046797 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657"} err="failed to get container status \"a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657\": rpc error: code = NotFound desc = could not find container \"a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657\": container with ID starting with a77a38287abb2df1912bfb48e0448e39f5fcafa933ec9557566a73034facc657 not found: ID does not exist" Oct 01 16:15:11 crc kubenswrapper[4688]: I1001 16:15:11.391666 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" path="/var/lib/kubelet/pods/a7e0b219-1147-4212-9a01-2d66a24601d7/volumes" Oct 01 16:15:16 crc kubenswrapper[4688]: I1001 16:15:16.573161 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:16 crc kubenswrapper[4688]: I1001 16:15:16.626320 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:17 crc kubenswrapper[4688]: I1001 16:15:17.392344 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:15:17 crc kubenswrapper[4688]: I1001 16:15:17.975664 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7t2zl" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="registry-server" containerID="cri-o://ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586" gracePeriod=2 Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.433339 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.622871 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities\") pod \"03fbc62e-54ec-447e-bc12-f50c317e6358\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.623288 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content\") pod \"03fbc62e-54ec-447e-bc12-f50c317e6358\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.623516 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6v5k\" (UniqueName: \"kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k\") pod \"03fbc62e-54ec-447e-bc12-f50c317e6358\" (UID: \"03fbc62e-54ec-447e-bc12-f50c317e6358\") " Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.623599 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities" (OuterVolumeSpecName: "utilities") pod "03fbc62e-54ec-447e-bc12-f50c317e6358" (UID: "03fbc62e-54ec-447e-bc12-f50c317e6358"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.624100 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.629511 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k" (OuterVolumeSpecName: "kube-api-access-f6v5k") pod "03fbc62e-54ec-447e-bc12-f50c317e6358" (UID: "03fbc62e-54ec-447e-bc12-f50c317e6358"). InnerVolumeSpecName "kube-api-access-f6v5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.707506 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03fbc62e-54ec-447e-bc12-f50c317e6358" (UID: "03fbc62e-54ec-447e-bc12-f50c317e6358"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.725449 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03fbc62e-54ec-447e-bc12-f50c317e6358-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.725489 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6v5k\" (UniqueName: \"kubernetes.io/projected/03fbc62e-54ec-447e-bc12-f50c317e6358-kube-api-access-f6v5k\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.985969 4688 generic.go:334] "Generic (PLEG): container finished" podID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerID="ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586" exitCode=0 Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.986058 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t2zl" Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.986067 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerDied","Data":"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586"} Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.988208 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t2zl" event={"ID":"03fbc62e-54ec-447e-bc12-f50c317e6358","Type":"ContainerDied","Data":"69b414e51549264efa64e6385b497486e852742480a08c8bfad7b9cf12968027"} Oct 01 16:15:18 crc kubenswrapper[4688]: I1001 16:15:18.988244 4688 scope.go:117] "RemoveContainer" containerID="ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.011786 4688 scope.go:117] "RemoveContainer" containerID="ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.027269 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.038511 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7t2zl"] Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.042200 4688 scope.go:117] "RemoveContainer" containerID="fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.098480 4688 scope.go:117] "RemoveContainer" containerID="ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586" Oct 01 16:15:19 crc kubenswrapper[4688]: E1001 16:15:19.098913 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586\": container with ID starting with ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586 not found: ID does not exist" containerID="ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.098949 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586"} err="failed to get container status \"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586\": rpc error: code = NotFound desc = could not find container \"ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586\": container with ID starting with ff31d90a71503645decf504362ac84f107f5c7576e432bcfe938f82191969586 not found: ID does not exist" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.098991 4688 scope.go:117] "RemoveContainer" containerID="ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3" Oct 01 16:15:19 crc kubenswrapper[4688]: E1001 16:15:19.099234 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3\": container with ID starting with ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3 not found: ID does not exist" containerID="ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.099253 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3"} err="failed to get container status \"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3\": rpc error: code = NotFound desc = could not find container \"ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3\": container with ID starting with ca303190a8184d8aa63be29fbd650993788f7930d45cf44c775e38d8aa5713a3 not found: ID does not exist" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.099267 4688 scope.go:117] "RemoveContainer" containerID="fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958" Oct 01 16:15:19 crc kubenswrapper[4688]: E1001 16:15:19.099481 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958\": container with ID starting with fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958 not found: ID does not exist" containerID="fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.099509 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958"} err="failed to get container status \"fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958\": rpc error: code = NotFound desc = could not find container \"fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958\": container with ID starting with fddf0ffa148f618bc6bc79efdc291a97945f23bf74f0806f869e9fecb6f11958 not found: ID does not exist" Oct 01 16:15:19 crc kubenswrapper[4688]: I1001 16:15:19.397484 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" path="/var/lib/kubelet/pods/03fbc62e-54ec-447e-bc12-f50c317e6358/volumes" Oct 01 16:15:21 crc kubenswrapper[4688]: I1001 16:15:21.410723 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:15:21 crc kubenswrapper[4688]: E1001 16:15:21.411394 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:15:32 crc kubenswrapper[4688]: I1001 16:15:32.381773 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:15:32 crc kubenswrapper[4688]: E1001 16:15:32.384401 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:15:39 crc kubenswrapper[4688]: I1001 16:15:39.166162 4688 generic.go:334] "Generic (PLEG): container finished" podID="9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" containerID="b54c3be600e557fc5a143f63318ad4f0fd8e199d5d46cb25c31f7e0ba41d6317" exitCode=0 Oct 01 16:15:39 crc kubenswrapper[4688]: I1001 16:15:39.166259 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" event={"ID":"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe","Type":"ContainerDied","Data":"b54c3be600e557fc5a143f63318ad4f0fd8e199d5d46cb25c31f7e0ba41d6317"} Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.560053 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.616936 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk8nm\" (UniqueName: \"kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm\") pod \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.617091 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key\") pod \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.617216 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory\") pod \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\" (UID: \"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe\") " Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.622311 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm" (OuterVolumeSpecName: "kube-api-access-jk8nm") pod "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" (UID: "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe"). InnerVolumeSpecName "kube-api-access-jk8nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.652692 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" (UID: "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.655294 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory" (OuterVolumeSpecName: "inventory") pod "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" (UID: "9cd2ef48-3ade-4edb-bd61-c3449a9c64fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.719619 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.719653 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk8nm\" (UniqueName: \"kubernetes.io/projected/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-kube-api-access-jk8nm\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:40 crc kubenswrapper[4688]: I1001 16:15:40.719664 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cd2ef48-3ade-4edb-bd61-c3449a9c64fe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.190149 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" event={"ID":"9cd2ef48-3ade-4edb-bd61-c3449a9c64fe","Type":"ContainerDied","Data":"0bb1b85b42a93915c2a4dfb20342d8b8ce75a4012bbffc634b27ba1b47d036e3"} Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.190212 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.190241 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bb1b85b42a93915c2a4dfb20342d8b8ce75a4012bbffc634b27ba1b47d036e3" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328254 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch"] Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328703 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="extract-utilities" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328717 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="extract-utilities" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328735 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328743 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328762 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="extract-utilities" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328771 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="extract-utilities" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328803 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="extract-content" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328811 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="extract-content" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328830 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328838 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328848 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328857 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328875 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="extract-content" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328883 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="extract-content" Oct 01 16:15:41 crc kubenswrapper[4688]: E1001 16:15:41.328895 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3c1bd4-a0a9-4778-8797-45bf6d6d017d" containerName="collect-profiles" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.328904 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3c1bd4-a0a9-4778-8797-45bf6d6d017d" containerName="collect-profiles" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.329113 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="03fbc62e-54ec-447e-bc12-f50c317e6358" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.329134 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3c1bd4-a0a9-4778-8797-45bf6d6d017d" containerName="collect-profiles" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.329147 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd2ef48-3ade-4edb-bd61-c3449a9c64fe" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.329174 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e0b219-1147-4212-9a01-2d66a24601d7" containerName="registry-server" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.329917 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.337256 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.337457 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.337557 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.338408 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.372114 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch"] Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.434291 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.434397 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chbbs\" (UniqueName: \"kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.434437 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.536156 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chbbs\" (UniqueName: \"kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.536295 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.536620 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.543212 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.544342 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.567938 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chbbs\" (UniqueName: \"kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k4lch\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:41 crc kubenswrapper[4688]: I1001 16:15:41.650778 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:15:42 crc kubenswrapper[4688]: I1001 16:15:42.201392 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch"] Oct 01 16:15:43 crc kubenswrapper[4688]: I1001 16:15:43.209210 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" event={"ID":"3fb9e457-e82c-4418-a5a6-a3892a359683","Type":"ContainerStarted","Data":"1ef670657ec5916a8c2bc0f52e2e623b565a40812ed5e9e8cbc1e34c44fa039b"} Oct 01 16:15:43 crc kubenswrapper[4688]: I1001 16:15:43.209514 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" event={"ID":"3fb9e457-e82c-4418-a5a6-a3892a359683","Type":"ContainerStarted","Data":"845348d0bdbfa97a82255184a24c7bc12eb1a3097171d85b70b80d65870b5582"} Oct 01 16:15:43 crc kubenswrapper[4688]: I1001 16:15:43.229411 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" podStartSLOduration=1.794205474 podStartE2EDuration="2.229379927s" podCreationTimestamp="2025-10-01 16:15:41 +0000 UTC" firstStartedPulling="2025-10-01 16:15:42.211293124 +0000 UTC m=+1731.561933086" lastFinishedPulling="2025-10-01 16:15:42.646467577 +0000 UTC m=+1731.997107539" observedRunningTime="2025-10-01 16:15:43.22450656 +0000 UTC m=+1732.575146532" watchObservedRunningTime="2025-10-01 16:15:43.229379927 +0000 UTC m=+1732.580019929" Oct 01 16:15:43 crc kubenswrapper[4688]: I1001 16:15:43.382402 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:15:43 crc kubenswrapper[4688]: E1001 16:15:43.382670 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:15:51 crc kubenswrapper[4688]: I1001 16:15:51.051728 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-df88w"] Oct 01 16:15:51 crc kubenswrapper[4688]: I1001 16:15:51.082683 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-df88w"] Oct 01 16:15:51 crc kubenswrapper[4688]: I1001 16:15:51.394768 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8dd8f9e-3618-40a4-a5a8-379142afdf65" path="/var/lib/kubelet/pods/c8dd8f9e-3618-40a4-a5a8-379142afdf65/volumes" Oct 01 16:15:52 crc kubenswrapper[4688]: I1001 16:15:52.022850 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-hdr4l"] Oct 01 16:15:52 crc kubenswrapper[4688]: I1001 16:15:52.029790 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-hdr4l"] Oct 01 16:15:53 crc kubenswrapper[4688]: I1001 16:15:53.031052 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-dn99k"] Oct 01 16:15:53 crc kubenswrapper[4688]: I1001 16:15:53.040718 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-dn99k"] Oct 01 16:15:53 crc kubenswrapper[4688]: I1001 16:15:53.395532 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf" path="/var/lib/kubelet/pods/af4d0e63-24aa-4f1a-b677-34d4fb3bc1bf/volumes" Oct 01 16:15:53 crc kubenswrapper[4688]: I1001 16:15:53.396087 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89bebef-c272-4141-9659-0621a89e1baa" path="/var/lib/kubelet/pods/f89bebef-c272-4141-9659-0621a89e1baa/volumes" Oct 01 16:15:57 crc kubenswrapper[4688]: I1001 16:15:57.382007 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:15:57 crc kubenswrapper[4688]: E1001 16:15:57.383233 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:16:00 crc kubenswrapper[4688]: I1001 16:16:00.686349 4688 scope.go:117] "RemoveContainer" containerID="d733f7af370aed949f3e5ed98c0c6853d1d7bf6928bf535783077588aaafb243" Oct 01 16:16:00 crc kubenswrapper[4688]: I1001 16:16:00.710719 4688 scope.go:117] "RemoveContainer" containerID="be433fced6fadcf1fae67dad74c865a4e88c858e8b4d13d20623521c4a9fdf47" Oct 01 16:16:00 crc kubenswrapper[4688]: I1001 16:16:00.766596 4688 scope.go:117] "RemoveContainer" containerID="d70829f518e96ec873ead8353a810739e03111c685e2196b634ec07a720ded11" Oct 01 16:16:00 crc kubenswrapper[4688]: I1001 16:16:00.809777 4688 scope.go:117] "RemoveContainer" containerID="26c7cb4e6b855793e97c207822f4a551fbb42b3d8e733269b88b27fcb86356b0" Oct 01 16:16:00 crc kubenswrapper[4688]: I1001 16:16:00.839409 4688 scope.go:117] "RemoveContainer" containerID="287025f6387c3e20194c2bf8af4a31a4fe8ed192538b47574fd21a7d50bfdff3" Oct 01 16:16:07 crc kubenswrapper[4688]: I1001 16:16:07.034002 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7327-account-create-7xldq"] Oct 01 16:16:07 crc kubenswrapper[4688]: I1001 16:16:07.043324 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7327-account-create-7xldq"] Oct 01 16:16:07 crc kubenswrapper[4688]: I1001 16:16:07.395619 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe56a6d-2d84-449b-b5a8-116c5a9891b1" path="/var/lib/kubelet/pods/cfe56a6d-2d84-449b-b5a8-116c5a9891b1/volumes" Oct 01 16:16:08 crc kubenswrapper[4688]: I1001 16:16:08.047894 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9cc9-account-create-pxjr8"] Oct 01 16:16:08 crc kubenswrapper[4688]: I1001 16:16:08.059447 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-71b5-account-create-v7977"] Oct 01 16:16:08 crc kubenswrapper[4688]: I1001 16:16:08.078795 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-71b5-account-create-v7977"] Oct 01 16:16:08 crc kubenswrapper[4688]: I1001 16:16:08.082146 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9cc9-account-create-pxjr8"] Oct 01 16:16:09 crc kubenswrapper[4688]: I1001 16:16:09.397352 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c68175-ca4c-4060-ac38-59f953323712" path="/var/lib/kubelet/pods/13c68175-ca4c-4060-ac38-59f953323712/volumes" Oct 01 16:16:09 crc kubenswrapper[4688]: I1001 16:16:09.399328 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635e82b7-cdaf-4cc7-b1a9-6ed165196067" path="/var/lib/kubelet/pods/635e82b7-cdaf-4cc7-b1a9-6ed165196067/volumes" Oct 01 16:16:10 crc kubenswrapper[4688]: I1001 16:16:10.381452 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:16:10 crc kubenswrapper[4688]: E1001 16:16:10.382230 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:16:22 crc kubenswrapper[4688]: I1001 16:16:22.383328 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:16:22 crc kubenswrapper[4688]: E1001 16:16:22.384686 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:16:33 crc kubenswrapper[4688]: I1001 16:16:33.054961 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w8mn9"] Oct 01 16:16:33 crc kubenswrapper[4688]: I1001 16:16:33.075615 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w8mn9"] Oct 01 16:16:33 crc kubenswrapper[4688]: I1001 16:16:33.407482 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7aa288b-66a4-423d-8e2c-2f1af84e429a" path="/var/lib/kubelet/pods/d7aa288b-66a4-423d-8e2c-2f1af84e429a/volumes" Oct 01 16:16:35 crc kubenswrapper[4688]: I1001 16:16:35.381592 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:16:35 crc kubenswrapper[4688]: E1001 16:16:35.381907 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:16:49 crc kubenswrapper[4688]: I1001 16:16:49.381212 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:16:49 crc kubenswrapper[4688]: E1001 16:16:49.381836 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:16:57 crc kubenswrapper[4688]: I1001 16:16:57.054000 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ds7rw"] Oct 01 16:16:57 crc kubenswrapper[4688]: I1001 16:16:57.061013 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ds7rw"] Oct 01 16:16:57 crc kubenswrapper[4688]: I1001 16:16:57.392935 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131a8df6-8ea9-48d0-a8e3-db8bec03dbcb" path="/var/lib/kubelet/pods/131a8df6-8ea9-48d0-a8e3-db8bec03dbcb/volumes" Oct 01 16:16:59 crc kubenswrapper[4688]: I1001 16:16:59.028913 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5xckd"] Oct 01 16:16:59 crc kubenswrapper[4688]: I1001 16:16:59.036995 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5xckd"] Oct 01 16:16:59 crc kubenswrapper[4688]: I1001 16:16:59.394854 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1" path="/var/lib/kubelet/pods/3cc5b93e-7d2d-4406-a5d0-8d1881cf54c1/volumes" Oct 01 16:17:00 crc kubenswrapper[4688]: I1001 16:17:00.381153 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.019069 4688 generic.go:334] "Generic (PLEG): container finished" podID="3fb9e457-e82c-4418-a5a6-a3892a359683" containerID="1ef670657ec5916a8c2bc0f52e2e623b565a40812ed5e9e8cbc1e34c44fa039b" exitCode=0 Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.019186 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" event={"ID":"3fb9e457-e82c-4418-a5a6-a3892a359683","Type":"ContainerDied","Data":"1ef670657ec5916a8c2bc0f52e2e623b565a40812ed5e9e8cbc1e34c44fa039b"} Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.023386 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f"} Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.053700 4688 scope.go:117] "RemoveContainer" containerID="39a07c3ed94706947f921a9d90921b151986bc3d04e5a090e52081b6183353f8" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.107834 4688 scope.go:117] "RemoveContainer" containerID="f9087ea8ce0a897ec369432bf53dcca1b4b86a133e060a4d45f85249cb82554f" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.156394 4688 scope.go:117] "RemoveContainer" containerID="98b4a26ee7f4613dbd0e313f650a37d1cdb03864dd603b409f71aa9c8b587ffa" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.202437 4688 scope.go:117] "RemoveContainer" containerID="e4dac6c5577c30afe60971b6dd2e7ec7a374ffeec22889c8ea8806dba1a45c21" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.258818 4688 scope.go:117] "RemoveContainer" containerID="0f04e8d5b0e7873574e92569f8b2f8367aa249018378f57c6b9fbc0f11c565a2" Oct 01 16:17:01 crc kubenswrapper[4688]: I1001 16:17:01.327390 4688 scope.go:117] "RemoveContainer" containerID="e6b15d46587538238a5d3597e85aa02d88c3f5042165db6bc8c89142e9e8120b" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.456844 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.533310 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chbbs\" (UniqueName: \"kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs\") pod \"3fb9e457-e82c-4418-a5a6-a3892a359683\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.533451 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key\") pod \"3fb9e457-e82c-4418-a5a6-a3892a359683\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.541430 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs" (OuterVolumeSpecName: "kube-api-access-chbbs") pod "3fb9e457-e82c-4418-a5a6-a3892a359683" (UID: "3fb9e457-e82c-4418-a5a6-a3892a359683"). InnerVolumeSpecName "kube-api-access-chbbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.567765 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3fb9e457-e82c-4418-a5a6-a3892a359683" (UID: "3fb9e457-e82c-4418-a5a6-a3892a359683"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.635848 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory\") pod \"3fb9e457-e82c-4418-a5a6-a3892a359683\" (UID: \"3fb9e457-e82c-4418-a5a6-a3892a359683\") " Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.636622 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chbbs\" (UniqueName: \"kubernetes.io/projected/3fb9e457-e82c-4418-a5a6-a3892a359683-kube-api-access-chbbs\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.636658 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.672355 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory" (OuterVolumeSpecName: "inventory") pod "3fb9e457-e82c-4418-a5a6-a3892a359683" (UID: "3fb9e457-e82c-4418-a5a6-a3892a359683"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:02 crc kubenswrapper[4688]: I1001 16:17:02.738251 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3fb9e457-e82c-4418-a5a6-a3892a359683-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.047499 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" event={"ID":"3fb9e457-e82c-4418-a5a6-a3892a359683","Type":"ContainerDied","Data":"845348d0bdbfa97a82255184a24c7bc12eb1a3097171d85b70b80d65870b5582"} Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.047579 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="845348d0bdbfa97a82255184a24c7bc12eb1a3097171d85b70b80d65870b5582" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.047639 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k4lch" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.145418 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p"] Oct 01 16:17:03 crc kubenswrapper[4688]: E1001 16:17:03.145941 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb9e457-e82c-4418-a5a6-a3892a359683" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.145965 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb9e457-e82c-4418-a5a6-a3892a359683" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.146247 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb9e457-e82c-4418-a5a6-a3892a359683" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.147132 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.152817 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.152896 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.153134 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.152983 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.157050 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p"] Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.250546 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln6sv\" (UniqueName: \"kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.250704 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.250753 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.352594 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln6sv\" (UniqueName: \"kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.352673 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.352706 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.362510 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.372931 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.397585 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln6sv\" (UniqueName: \"kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-whw8p\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:03 crc kubenswrapper[4688]: I1001 16:17:03.480114 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:04 crc kubenswrapper[4688]: I1001 16:17:04.015344 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p"] Oct 01 16:17:04 crc kubenswrapper[4688]: I1001 16:17:04.059612 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" event={"ID":"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862","Type":"ContainerStarted","Data":"dddbf3808a9aee041c915f4889d8066facbec7525f88f7f87988c48228db6be9"} Oct 01 16:17:05 crc kubenswrapper[4688]: I1001 16:17:05.068595 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" event={"ID":"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862","Type":"ContainerStarted","Data":"5cca8d7e6c84461ae273588504255b7bb5f730f65d0da8257bb66f8b600dd37c"} Oct 01 16:17:05 crc kubenswrapper[4688]: I1001 16:17:05.089963 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" podStartSLOduration=1.664835289 podStartE2EDuration="2.089945651s" podCreationTimestamp="2025-10-01 16:17:03 +0000 UTC" firstStartedPulling="2025-10-01 16:17:04.02640263 +0000 UTC m=+1813.377042592" lastFinishedPulling="2025-10-01 16:17:04.451512992 +0000 UTC m=+1813.802152954" observedRunningTime="2025-10-01 16:17:05.084067274 +0000 UTC m=+1814.434707236" watchObservedRunningTime="2025-10-01 16:17:05.089945651 +0000 UTC m=+1814.440585613" Oct 01 16:17:10 crc kubenswrapper[4688]: I1001 16:17:10.117986 4688 generic.go:334] "Generic (PLEG): container finished" podID="fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" containerID="5cca8d7e6c84461ae273588504255b7bb5f730f65d0da8257bb66f8b600dd37c" exitCode=0 Oct 01 16:17:10 crc kubenswrapper[4688]: I1001 16:17:10.118052 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" event={"ID":"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862","Type":"ContainerDied","Data":"5cca8d7e6c84461ae273588504255b7bb5f730f65d0da8257bb66f8b600dd37c"} Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.560131 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.631665 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key\") pod \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.631970 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln6sv\" (UniqueName: \"kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv\") pod \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.632026 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory\") pod \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\" (UID: \"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862\") " Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.640721 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv" (OuterVolumeSpecName: "kube-api-access-ln6sv") pod "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" (UID: "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862"). InnerVolumeSpecName "kube-api-access-ln6sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.669456 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory" (OuterVolumeSpecName: "inventory") pod "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" (UID: "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.706069 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" (UID: "fd4cc96c-61fa-4f47-8a36-8d2c6ddef862"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.709582 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs"] Oct 01 16:17:11 crc kubenswrapper[4688]: E1001 16:17:11.710006 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.710022 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.710217 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd4cc96c-61fa-4f47-8a36-8d2c6ddef862" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.711229 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.723186 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs"] Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.772744 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln6sv\" (UniqueName: \"kubernetes.io/projected/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-kube-api-access-ln6sv\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.772768 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.772778 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4cc96c-61fa-4f47-8a36-8d2c6ddef862-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.874997 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.875482 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.875598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stxzt\" (UniqueName: \"kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.977416 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.977472 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stxzt\" (UniqueName: \"kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.977545 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.982598 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:11 crc kubenswrapper[4688]: I1001 16:17:11.982627 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.007911 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stxzt\" (UniqueName: \"kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdvcs\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.100493 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.140629 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" event={"ID":"fd4cc96c-61fa-4f47-8a36-8d2c6ddef862","Type":"ContainerDied","Data":"dddbf3808a9aee041c915f4889d8066facbec7525f88f7f87988c48228db6be9"} Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.140688 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dddbf3808a9aee041c915f4889d8066facbec7525f88f7f87988c48228db6be9" Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.140764 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-whw8p" Oct 01 16:17:12 crc kubenswrapper[4688]: I1001 16:17:12.619304 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs"] Oct 01 16:17:13 crc kubenswrapper[4688]: I1001 16:17:13.150372 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" event={"ID":"77d2978e-28f0-4591-a3fc-491ebfb6d1d9","Type":"ContainerStarted","Data":"a9d3870db8e5a00cc7e33bdd5c4145cedf6eee535041e602005ea9f1a57de44e"} Oct 01 16:17:14 crc kubenswrapper[4688]: I1001 16:17:14.161944 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" event={"ID":"77d2978e-28f0-4591-a3fc-491ebfb6d1d9","Type":"ContainerStarted","Data":"2deeb52c107835f95bd8a71ae489a4665e374ac9d24b7ab7fd308a4f0315ebfb"} Oct 01 16:17:14 crc kubenswrapper[4688]: I1001 16:17:14.180869 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" podStartSLOduration=2.56752903 podStartE2EDuration="3.180847885s" podCreationTimestamp="2025-10-01 16:17:11 +0000 UTC" firstStartedPulling="2025-10-01 16:17:12.624129821 +0000 UTC m=+1821.974769793" lastFinishedPulling="2025-10-01 16:17:13.237448686 +0000 UTC m=+1822.588088648" observedRunningTime="2025-10-01 16:17:14.178265741 +0000 UTC m=+1823.528905713" watchObservedRunningTime="2025-10-01 16:17:14.180847885 +0000 UTC m=+1823.531487867" Oct 01 16:17:44 crc kubenswrapper[4688]: I1001 16:17:44.049858 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbx2w"] Oct 01 16:17:44 crc kubenswrapper[4688]: I1001 16:17:44.063540 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rbx2w"] Oct 01 16:17:45 crc kubenswrapper[4688]: I1001 16:17:45.391913 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8199e7d8-c362-43ae-a734-d287ff9f293d" path="/var/lib/kubelet/pods/8199e7d8-c362-43ae-a734-d287ff9f293d/volumes" Oct 01 16:17:52 crc kubenswrapper[4688]: I1001 16:17:52.529061 4688 generic.go:334] "Generic (PLEG): container finished" podID="77d2978e-28f0-4591-a3fc-491ebfb6d1d9" containerID="2deeb52c107835f95bd8a71ae489a4665e374ac9d24b7ab7fd308a4f0315ebfb" exitCode=0 Oct 01 16:17:52 crc kubenswrapper[4688]: I1001 16:17:52.529510 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" event={"ID":"77d2978e-28f0-4591-a3fc-491ebfb6d1d9","Type":"ContainerDied","Data":"2deeb52c107835f95bd8a71ae489a4665e374ac9d24b7ab7fd308a4f0315ebfb"} Oct 01 16:17:53 crc kubenswrapper[4688]: I1001 16:17:53.990004 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.158660 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stxzt\" (UniqueName: \"kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt\") pod \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.158728 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key\") pod \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.158754 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory\") pod \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\" (UID: \"77d2978e-28f0-4591-a3fc-491ebfb6d1d9\") " Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.168596 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt" (OuterVolumeSpecName: "kube-api-access-stxzt") pod "77d2978e-28f0-4591-a3fc-491ebfb6d1d9" (UID: "77d2978e-28f0-4591-a3fc-491ebfb6d1d9"). InnerVolumeSpecName "kube-api-access-stxzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.195190 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory" (OuterVolumeSpecName: "inventory") pod "77d2978e-28f0-4591-a3fc-491ebfb6d1d9" (UID: "77d2978e-28f0-4591-a3fc-491ebfb6d1d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.212720 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77d2978e-28f0-4591-a3fc-491ebfb6d1d9" (UID: "77d2978e-28f0-4591-a3fc-491ebfb6d1d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.260935 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stxzt\" (UniqueName: \"kubernetes.io/projected/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-kube-api-access-stxzt\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.260970 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.260979 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77d2978e-28f0-4591-a3fc-491ebfb6d1d9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.554593 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" event={"ID":"77d2978e-28f0-4591-a3fc-491ebfb6d1d9","Type":"ContainerDied","Data":"a9d3870db8e5a00cc7e33bdd5c4145cedf6eee535041e602005ea9f1a57de44e"} Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.555037 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9d3870db8e5a00cc7e33bdd5c4145cedf6eee535041e602005ea9f1a57de44e" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.554693 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdvcs" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.683063 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk"] Oct 01 16:17:54 crc kubenswrapper[4688]: E1001 16:17:54.683856 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77d2978e-28f0-4591-a3fc-491ebfb6d1d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.683882 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="77d2978e-28f0-4591-a3fc-491ebfb6d1d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.684099 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="77d2978e-28f0-4591-a3fc-491ebfb6d1d9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.685040 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.686946 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.687408 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.687661 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.688565 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.711258 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk"] Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.774844 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hzlz\" (UniqueName: \"kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.775213 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.775422 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.882184 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hzlz\" (UniqueName: \"kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.882293 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.882339 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.893350 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.924359 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:54 crc kubenswrapper[4688]: I1001 16:17:54.943730 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hzlz\" (UniqueName: \"kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:55 crc kubenswrapper[4688]: I1001 16:17:55.004006 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:17:55 crc kubenswrapper[4688]: I1001 16:17:55.602179 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk"] Oct 01 16:17:55 crc kubenswrapper[4688]: W1001 16:17:55.610668 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4633912c_f539_44fd_8fc4_1e57ddbdad56.slice/crio-5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d WatchSource:0}: Error finding container 5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d: Status 404 returned error can't find the container with id 5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d Oct 01 16:17:56 crc kubenswrapper[4688]: I1001 16:17:56.573042 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" event={"ID":"4633912c-f539-44fd-8fc4-1e57ddbdad56","Type":"ContainerStarted","Data":"f085784328f005f9e9c566af644f954b517215338bd9568071271f0ede877dcb"} Oct 01 16:17:56 crc kubenswrapper[4688]: I1001 16:17:56.573375 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" event={"ID":"4633912c-f539-44fd-8fc4-1e57ddbdad56","Type":"ContainerStarted","Data":"5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d"} Oct 01 16:17:56 crc kubenswrapper[4688]: I1001 16:17:56.592425 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" podStartSLOduration=1.9955496940000002 podStartE2EDuration="2.59240326s" podCreationTimestamp="2025-10-01 16:17:54 +0000 UTC" firstStartedPulling="2025-10-01 16:17:55.613111438 +0000 UTC m=+1864.963751400" lastFinishedPulling="2025-10-01 16:17:56.209965004 +0000 UTC m=+1865.560604966" observedRunningTime="2025-10-01 16:17:56.586023758 +0000 UTC m=+1865.936663720" watchObservedRunningTime="2025-10-01 16:17:56.59240326 +0000 UTC m=+1865.943043232" Oct 01 16:18:01 crc kubenswrapper[4688]: I1001 16:18:01.450122 4688 scope.go:117] "RemoveContainer" containerID="86e31a9a079694bb042db75db204c2f0d3ede0e67af2ff1083533099cfa214dd" Oct 01 16:18:53 crc kubenswrapper[4688]: I1001 16:18:53.086077 4688 generic.go:334] "Generic (PLEG): container finished" podID="4633912c-f539-44fd-8fc4-1e57ddbdad56" containerID="f085784328f005f9e9c566af644f954b517215338bd9568071271f0ede877dcb" exitCode=2 Oct 01 16:18:53 crc kubenswrapper[4688]: I1001 16:18:53.086677 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" event={"ID":"4633912c-f539-44fd-8fc4-1e57ddbdad56","Type":"ContainerDied","Data":"f085784328f005f9e9c566af644f954b517215338bd9568071271f0ede877dcb"} Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.507540 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.610284 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory\") pod \"4633912c-f539-44fd-8fc4-1e57ddbdad56\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.610998 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key\") pod \"4633912c-f539-44fd-8fc4-1e57ddbdad56\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.611183 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hzlz\" (UniqueName: \"kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz\") pod \"4633912c-f539-44fd-8fc4-1e57ddbdad56\" (UID: \"4633912c-f539-44fd-8fc4-1e57ddbdad56\") " Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.615722 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz" (OuterVolumeSpecName: "kube-api-access-7hzlz") pod "4633912c-f539-44fd-8fc4-1e57ddbdad56" (UID: "4633912c-f539-44fd-8fc4-1e57ddbdad56"). InnerVolumeSpecName "kube-api-access-7hzlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.641157 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4633912c-f539-44fd-8fc4-1e57ddbdad56" (UID: "4633912c-f539-44fd-8fc4-1e57ddbdad56"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.651746 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory" (OuterVolumeSpecName: "inventory") pod "4633912c-f539-44fd-8fc4-1e57ddbdad56" (UID: "4633912c-f539-44fd-8fc4-1e57ddbdad56"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.713021 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.713064 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hzlz\" (UniqueName: \"kubernetes.io/projected/4633912c-f539-44fd-8fc4-1e57ddbdad56-kube-api-access-7hzlz\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:54 crc kubenswrapper[4688]: I1001 16:18:54.713076 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4633912c-f539-44fd-8fc4-1e57ddbdad56-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:18:55 crc kubenswrapper[4688]: I1001 16:18:55.103452 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" event={"ID":"4633912c-f539-44fd-8fc4-1e57ddbdad56","Type":"ContainerDied","Data":"5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d"} Oct 01 16:18:55 crc kubenswrapper[4688]: I1001 16:18:55.103493 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5650244e7ed7838d0eeb084576a7cbbe921b42c5847c46fb9e125201c24feb8d" Oct 01 16:18:55 crc kubenswrapper[4688]: I1001 16:18:55.103502 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.027052 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6"] Oct 01 16:19:02 crc kubenswrapper[4688]: E1001 16:19:02.028032 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4633912c-f539-44fd-8fc4-1e57ddbdad56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.028047 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="4633912c-f539-44fd-8fc4-1e57ddbdad56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.028239 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="4633912c-f539-44fd-8fc4-1e57ddbdad56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.028879 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.031202 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.035629 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.035649 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.036176 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.039338 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6"] Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.150613 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.150770 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.151424 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx5v4\" (UniqueName: \"kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.253915 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.254074 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx5v4\" (UniqueName: \"kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.254140 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.261465 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.263405 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.284303 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx5v4\" (UniqueName: \"kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.348704 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.932774 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6"] Oct 01 16:19:02 crc kubenswrapper[4688]: I1001 16:19:02.940399 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:19:03 crc kubenswrapper[4688]: I1001 16:19:03.173141 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" event={"ID":"42f771c8-2902-4358-8f27-3dd4e9ac510e","Type":"ContainerStarted","Data":"7cea8f8cee032a14342b7d97b80971846addcb790b24aa853ce09a291b5b3551"} Oct 01 16:19:04 crc kubenswrapper[4688]: I1001 16:19:04.187155 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" event={"ID":"42f771c8-2902-4358-8f27-3dd4e9ac510e","Type":"ContainerStarted","Data":"83253f0515b87351bd42b925b0eb30fc2cdca1a45c3f0f68be51684c76d3db19"} Oct 01 16:19:04 crc kubenswrapper[4688]: I1001 16:19:04.213493 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" podStartSLOduration=1.6722244910000001 podStartE2EDuration="2.213470502s" podCreationTimestamp="2025-10-01 16:19:02 +0000 UTC" firstStartedPulling="2025-10-01 16:19:02.940169293 +0000 UTC m=+1932.290809255" lastFinishedPulling="2025-10-01 16:19:03.481415304 +0000 UTC m=+1932.832055266" observedRunningTime="2025-10-01 16:19:04.203540909 +0000 UTC m=+1933.554180871" watchObservedRunningTime="2025-10-01 16:19:04.213470502 +0000 UTC m=+1933.564110474" Oct 01 16:19:25 crc kubenswrapper[4688]: I1001 16:19:25.053052 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:19:25 crc kubenswrapper[4688]: I1001 16:19:25.053628 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:19:49 crc kubenswrapper[4688]: I1001 16:19:49.567735 4688 generic.go:334] "Generic (PLEG): container finished" podID="42f771c8-2902-4358-8f27-3dd4e9ac510e" containerID="83253f0515b87351bd42b925b0eb30fc2cdca1a45c3f0f68be51684c76d3db19" exitCode=0 Oct 01 16:19:49 crc kubenswrapper[4688]: I1001 16:19:49.567926 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" event={"ID":"42f771c8-2902-4358-8f27-3dd4e9ac510e","Type":"ContainerDied","Data":"83253f0515b87351bd42b925b0eb30fc2cdca1a45c3f0f68be51684c76d3db19"} Oct 01 16:19:50 crc kubenswrapper[4688]: I1001 16:19:50.969145 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.113492 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory\") pod \"42f771c8-2902-4358-8f27-3dd4e9ac510e\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.113642 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx5v4\" (UniqueName: \"kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4\") pod \"42f771c8-2902-4358-8f27-3dd4e9ac510e\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.113867 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key\") pod \"42f771c8-2902-4358-8f27-3dd4e9ac510e\" (UID: \"42f771c8-2902-4358-8f27-3dd4e9ac510e\") " Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.118416 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4" (OuterVolumeSpecName: "kube-api-access-sx5v4") pod "42f771c8-2902-4358-8f27-3dd4e9ac510e" (UID: "42f771c8-2902-4358-8f27-3dd4e9ac510e"). InnerVolumeSpecName "kube-api-access-sx5v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.143462 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory" (OuterVolumeSpecName: "inventory") pod "42f771c8-2902-4358-8f27-3dd4e9ac510e" (UID: "42f771c8-2902-4358-8f27-3dd4e9ac510e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.152677 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "42f771c8-2902-4358-8f27-3dd4e9ac510e" (UID: "42f771c8-2902-4358-8f27-3dd4e9ac510e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.216182 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.216232 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx5v4\" (UniqueName: \"kubernetes.io/projected/42f771c8-2902-4358-8f27-3dd4e9ac510e-kube-api-access-sx5v4\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.216251 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42f771c8-2902-4358-8f27-3dd4e9ac510e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.586666 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" event={"ID":"42f771c8-2902-4358-8f27-3dd4e9ac510e","Type":"ContainerDied","Data":"7cea8f8cee032a14342b7d97b80971846addcb790b24aa853ce09a291b5b3551"} Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.586747 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cea8f8cee032a14342b7d97b80971846addcb790b24aa853ce09a291b5b3551" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.586701 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.693625 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bj9n4"] Oct 01 16:19:51 crc kubenswrapper[4688]: E1001 16:19:51.694125 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f771c8-2902-4358-8f27-3dd4e9ac510e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.694154 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f771c8-2902-4358-8f27-3dd4e9ac510e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.694374 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f771c8-2902-4358-8f27-3dd4e9ac510e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.695153 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.702128 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.703139 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.703310 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.705163 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.707169 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bj9n4"] Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.830225 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmjpc\" (UniqueName: \"kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.830499 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.830545 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.932211 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmjpc\" (UniqueName: \"kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.932314 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.932347 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.938561 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.948136 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:51 crc kubenswrapper[4688]: I1001 16:19:51.953302 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmjpc\" (UniqueName: \"kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc\") pod \"ssh-known-hosts-edpm-deployment-bj9n4\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:52 crc kubenswrapper[4688]: I1001 16:19:52.016989 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:19:52 crc kubenswrapper[4688]: I1001 16:19:52.564905 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-bj9n4"] Oct 01 16:19:52 crc kubenswrapper[4688]: I1001 16:19:52.597321 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" event={"ID":"289bca66-8942-4580-b92e-d5cb8db5deed","Type":"ContainerStarted","Data":"2f4f60380b854b27e30607c310b22d222b7abb06006a09fa6888bbb2a69cebde"} Oct 01 16:19:53 crc kubenswrapper[4688]: I1001 16:19:53.612637 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" event={"ID":"289bca66-8942-4580-b92e-d5cb8db5deed","Type":"ContainerStarted","Data":"ccdce4e98e0c261b81a9444d0f764a92930b863905c579ed2e164bbb9a461677"} Oct 01 16:19:55 crc kubenswrapper[4688]: I1001 16:19:55.052691 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:19:55 crc kubenswrapper[4688]: I1001 16:19:55.053068 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:20:00 crc kubenswrapper[4688]: I1001 16:20:00.686831 4688 generic.go:334] "Generic (PLEG): container finished" podID="289bca66-8942-4580-b92e-d5cb8db5deed" containerID="ccdce4e98e0c261b81a9444d0f764a92930b863905c579ed2e164bbb9a461677" exitCode=0 Oct 01 16:20:00 crc kubenswrapper[4688]: I1001 16:20:00.686933 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" event={"ID":"289bca66-8942-4580-b92e-d5cb8db5deed","Type":"ContainerDied","Data":"ccdce4e98e0c261b81a9444d0f764a92930b863905c579ed2e164bbb9a461677"} Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.147880 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.327718 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam\") pod \"289bca66-8942-4580-b92e-d5cb8db5deed\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.328302 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0\") pod \"289bca66-8942-4580-b92e-d5cb8db5deed\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.328400 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmjpc\" (UniqueName: \"kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc\") pod \"289bca66-8942-4580-b92e-d5cb8db5deed\" (UID: \"289bca66-8942-4580-b92e-d5cb8db5deed\") " Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.337496 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc" (OuterVolumeSpecName: "kube-api-access-qmjpc") pod "289bca66-8942-4580-b92e-d5cb8db5deed" (UID: "289bca66-8942-4580-b92e-d5cb8db5deed"). InnerVolumeSpecName "kube-api-access-qmjpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.360782 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "289bca66-8942-4580-b92e-d5cb8db5deed" (UID: "289bca66-8942-4580-b92e-d5cb8db5deed"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.369131 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "289bca66-8942-4580-b92e-d5cb8db5deed" (UID: "289bca66-8942-4580-b92e-d5cb8db5deed"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.430157 4688 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.430183 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmjpc\" (UniqueName: \"kubernetes.io/projected/289bca66-8942-4580-b92e-d5cb8db5deed-kube-api-access-qmjpc\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.430197 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/289bca66-8942-4580-b92e-d5cb8db5deed-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.709866 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" event={"ID":"289bca66-8942-4580-b92e-d5cb8db5deed","Type":"ContainerDied","Data":"2f4f60380b854b27e30607c310b22d222b7abb06006a09fa6888bbb2a69cebde"} Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.710224 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f4f60380b854b27e30607c310b22d222b7abb06006a09fa6888bbb2a69cebde" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.709923 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-bj9n4" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.797401 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9"] Oct 01 16:20:02 crc kubenswrapper[4688]: E1001 16:20:02.797947 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289bca66-8942-4580-b92e-d5cb8db5deed" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.797972 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="289bca66-8942-4580-b92e-d5cb8db5deed" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.798222 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="289bca66-8942-4580-b92e-d5cb8db5deed" containerName="ssh-known-hosts-edpm-deployment" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.798979 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.807640 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9"] Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.809116 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.809641 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.810044 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.810551 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.938192 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.938444 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:02 crc kubenswrapper[4688]: I1001 16:20:02.938663 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb49k\" (UniqueName: \"kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.040535 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb49k\" (UniqueName: \"kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.040636 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.040669 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.044941 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.048251 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.066358 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb49k\" (UniqueName: \"kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r7vc9\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.124063 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.646398 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9"] Oct 01 16:20:03 crc kubenswrapper[4688]: I1001 16:20:03.730238 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" event={"ID":"a1720cc4-8ce5-4d4a-b0f0-47afac314627","Type":"ContainerStarted","Data":"1e7c9167ca22a6ef4c1a13795429249ed3afbac63049703b9e85996b48b2ecb7"} Oct 01 16:20:04 crc kubenswrapper[4688]: I1001 16:20:04.739079 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" event={"ID":"a1720cc4-8ce5-4d4a-b0f0-47afac314627","Type":"ContainerStarted","Data":"d22b0997631471fb75c4bce7b56898aa971ac82fc9628faa9d75114984174e4a"} Oct 01 16:20:04 crc kubenswrapper[4688]: I1001 16:20:04.765683 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" podStartSLOduration=2.07281251 podStartE2EDuration="2.76566221s" podCreationTimestamp="2025-10-01 16:20:02 +0000 UTC" firstStartedPulling="2025-10-01 16:20:03.651380643 +0000 UTC m=+1993.002020605" lastFinishedPulling="2025-10-01 16:20:04.344230343 +0000 UTC m=+1993.694870305" observedRunningTime="2025-10-01 16:20:04.751998301 +0000 UTC m=+1994.102638263" watchObservedRunningTime="2025-10-01 16:20:04.76566221 +0000 UTC m=+1994.116302172" Oct 01 16:20:12 crc kubenswrapper[4688]: I1001 16:20:12.812169 4688 generic.go:334] "Generic (PLEG): container finished" podID="a1720cc4-8ce5-4d4a-b0f0-47afac314627" containerID="d22b0997631471fb75c4bce7b56898aa971ac82fc9628faa9d75114984174e4a" exitCode=0 Oct 01 16:20:12 crc kubenswrapper[4688]: I1001 16:20:12.812254 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" event={"ID":"a1720cc4-8ce5-4d4a-b0f0-47afac314627","Type":"ContainerDied","Data":"d22b0997631471fb75c4bce7b56898aa971ac82fc9628faa9d75114984174e4a"} Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.280993 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.469322 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key\") pod \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.469861 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory\") pod \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.469963 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb49k\" (UniqueName: \"kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k\") pod \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\" (UID: \"a1720cc4-8ce5-4d4a-b0f0-47afac314627\") " Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.479834 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k" (OuterVolumeSpecName: "kube-api-access-sb49k") pod "a1720cc4-8ce5-4d4a-b0f0-47afac314627" (UID: "a1720cc4-8ce5-4d4a-b0f0-47afac314627"). InnerVolumeSpecName "kube-api-access-sb49k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.499248 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1720cc4-8ce5-4d4a-b0f0-47afac314627" (UID: "a1720cc4-8ce5-4d4a-b0f0-47afac314627"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.499377 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory" (OuterVolumeSpecName: "inventory") pod "a1720cc4-8ce5-4d4a-b0f0-47afac314627" (UID: "a1720cc4-8ce5-4d4a-b0f0-47afac314627"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.573742 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.573785 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1720cc4-8ce5-4d4a-b0f0-47afac314627-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.573798 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb49k\" (UniqueName: \"kubernetes.io/projected/a1720cc4-8ce5-4d4a-b0f0-47afac314627-kube-api-access-sb49k\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.839817 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" event={"ID":"a1720cc4-8ce5-4d4a-b0f0-47afac314627","Type":"ContainerDied","Data":"1e7c9167ca22a6ef4c1a13795429249ed3afbac63049703b9e85996b48b2ecb7"} Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.839861 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e7c9167ca22a6ef4c1a13795429249ed3afbac63049703b9e85996b48b2ecb7" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.839923 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r7vc9" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.914891 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p"] Oct 01 16:20:14 crc kubenswrapper[4688]: E1001 16:20:14.915726 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1720cc4-8ce5-4d4a-b0f0-47afac314627" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.915752 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1720cc4-8ce5-4d4a-b0f0-47afac314627" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.916000 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1720cc4-8ce5-4d4a-b0f0-47afac314627" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.916772 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.919350 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.919546 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.919897 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.921836 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:20:14 crc kubenswrapper[4688]: I1001 16:20:14.928660 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p"] Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.082946 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.083029 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.083127 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knfrc\" (UniqueName: \"kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.185625 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.185808 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.185937 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knfrc\" (UniqueName: \"kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.198728 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.200231 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.203634 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knfrc\" (UniqueName: \"kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c755p\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.238263 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.809515 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p"] Oct 01 16:20:15 crc kubenswrapper[4688]: I1001 16:20:15.851729 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" event={"ID":"9cebafe1-24f0-4751-b691-8fd0bc09359f","Type":"ContainerStarted","Data":"fc7e1044654c5b58a050f37b871a1e465c24ce59449f9b873d3d25ddbe938ea4"} Oct 01 16:20:16 crc kubenswrapper[4688]: I1001 16:20:16.864207 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" event={"ID":"9cebafe1-24f0-4751-b691-8fd0bc09359f","Type":"ContainerStarted","Data":"a2c164935ab4bb5070980b855f72609883a2ab805f69f460ccc45b49fedce0cf"} Oct 01 16:20:16 crc kubenswrapper[4688]: I1001 16:20:16.886902 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" podStartSLOduration=2.488902313 podStartE2EDuration="2.886881342s" podCreationTimestamp="2025-10-01 16:20:14 +0000 UTC" firstStartedPulling="2025-10-01 16:20:15.827773586 +0000 UTC m=+2005.178413558" lastFinishedPulling="2025-10-01 16:20:16.225752625 +0000 UTC m=+2005.576392587" observedRunningTime="2025-10-01 16:20:16.881226861 +0000 UTC m=+2006.231866823" watchObservedRunningTime="2025-10-01 16:20:16.886881342 +0000 UTC m=+2006.237521304" Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.052991 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.053826 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.053910 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.055200 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.055299 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f" gracePeriod=600 Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.956504 4688 generic.go:334] "Generic (PLEG): container finished" podID="9cebafe1-24f0-4751-b691-8fd0bc09359f" containerID="a2c164935ab4bb5070980b855f72609883a2ab805f69f460ccc45b49fedce0cf" exitCode=0 Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.956745 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" event={"ID":"9cebafe1-24f0-4751-b691-8fd0bc09359f","Type":"ContainerDied","Data":"a2c164935ab4bb5070980b855f72609883a2ab805f69f460ccc45b49fedce0cf"} Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.963134 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f" exitCode=0 Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.963165 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f"} Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.963234 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b"} Oct 01 16:20:25 crc kubenswrapper[4688]: I1001 16:20:25.963272 4688 scope.go:117] "RemoveContainer" containerID="d282d57140015e6e0d856064a6867053eebb7915ee325d332266fbaac1ab85b7" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.394606 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.543034 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key\") pod \"9cebafe1-24f0-4751-b691-8fd0bc09359f\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.543328 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory\") pod \"9cebafe1-24f0-4751-b691-8fd0bc09359f\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.543510 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knfrc\" (UniqueName: \"kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc\") pod \"9cebafe1-24f0-4751-b691-8fd0bc09359f\" (UID: \"9cebafe1-24f0-4751-b691-8fd0bc09359f\") " Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.554884 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc" (OuterVolumeSpecName: "kube-api-access-knfrc") pod "9cebafe1-24f0-4751-b691-8fd0bc09359f" (UID: "9cebafe1-24f0-4751-b691-8fd0bc09359f"). InnerVolumeSpecName "kube-api-access-knfrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.568678 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9cebafe1-24f0-4751-b691-8fd0bc09359f" (UID: "9cebafe1-24f0-4751-b691-8fd0bc09359f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.578599 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory" (OuterVolumeSpecName: "inventory") pod "9cebafe1-24f0-4751-b691-8fd0bc09359f" (UID: "9cebafe1-24f0-4751-b691-8fd0bc09359f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.645483 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.645510 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cebafe1-24f0-4751-b691-8fd0bc09359f-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.645537 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knfrc\" (UniqueName: \"kubernetes.io/projected/9cebafe1-24f0-4751-b691-8fd0bc09359f-kube-api-access-knfrc\") on node \"crc\" DevicePath \"\"" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.986057 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" event={"ID":"9cebafe1-24f0-4751-b691-8fd0bc09359f","Type":"ContainerDied","Data":"fc7e1044654c5b58a050f37b871a1e465c24ce59449f9b873d3d25ddbe938ea4"} Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.986563 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc7e1044654c5b58a050f37b871a1e465c24ce59449f9b873d3d25ddbe938ea4" Oct 01 16:20:27 crc kubenswrapper[4688]: I1001 16:20:27.986100 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c755p" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.168991 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x"] Oct 01 16:20:28 crc kubenswrapper[4688]: E1001 16:20:28.169458 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cebafe1-24f0-4751-b691-8fd0bc09359f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.169478 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cebafe1-24f0-4751-b691-8fd0bc09359f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.169713 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cebafe1-24f0-4751-b691-8fd0bc09359f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.170500 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.176362 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.176374 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.176668 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.176437 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.176498 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.177008 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.177426 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.178746 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.190424 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x"] Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.256932 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257153 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257195 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257284 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257318 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257353 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257446 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257496 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257544 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257568 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257595 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257615 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mttsr\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.257638 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.358901 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.358975 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359011 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359039 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359086 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359116 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359143 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mttsr\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359168 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359199 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359249 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359303 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359343 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359412 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.359444 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.369846 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.371096 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.371650 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.372249 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.372640 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.373211 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.373623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.373624 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.374790 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.375214 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.376802 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.377734 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.383838 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.385752 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mttsr\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:28 crc kubenswrapper[4688]: I1001 16:20:28.492370 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:20:29 crc kubenswrapper[4688]: I1001 16:20:29.046151 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x"] Oct 01 16:20:30 crc kubenswrapper[4688]: I1001 16:20:30.004606 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" event={"ID":"944f4da8-c2bb-400d-a78f-caf11336d5ec","Type":"ContainerStarted","Data":"703dc78bc76459df48fb0970304560bccfc2937a0e3991e5358815cf4b6b4e9b"} Oct 01 16:20:30 crc kubenswrapper[4688]: I1001 16:20:30.005136 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" event={"ID":"944f4da8-c2bb-400d-a78f-caf11336d5ec","Type":"ContainerStarted","Data":"28debda9718821eff1cd04d127c40cbd8303b213bdd09df8e0c09ab7cb6bcd63"} Oct 01 16:20:30 crc kubenswrapper[4688]: I1001 16:20:30.027920 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" podStartSLOduration=1.5365325680000002 podStartE2EDuration="2.027899528s" podCreationTimestamp="2025-10-01 16:20:28 +0000 UTC" firstStartedPulling="2025-10-01 16:20:29.061349129 +0000 UTC m=+2018.411989091" lastFinishedPulling="2025-10-01 16:20:29.552716089 +0000 UTC m=+2018.903356051" observedRunningTime="2025-10-01 16:20:30.021930808 +0000 UTC m=+2019.372570790" watchObservedRunningTime="2025-10-01 16:20:30.027899528 +0000 UTC m=+2019.378539490" Oct 01 16:21:09 crc kubenswrapper[4688]: I1001 16:21:09.388451 4688 generic.go:334] "Generic (PLEG): container finished" podID="944f4da8-c2bb-400d-a78f-caf11336d5ec" containerID="703dc78bc76459df48fb0970304560bccfc2937a0e3991e5358815cf4b6b4e9b" exitCode=0 Oct 01 16:21:09 crc kubenswrapper[4688]: I1001 16:21:09.390509 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" event={"ID":"944f4da8-c2bb-400d-a78f-caf11336d5ec","Type":"ContainerDied","Data":"703dc78bc76459df48fb0970304560bccfc2937a0e3991e5358815cf4b6b4e9b"} Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.848289 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873067 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873122 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873163 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873189 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873219 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873259 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873307 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873343 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873388 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873417 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mttsr\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873453 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873497 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873547 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.873579 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle\") pod \"944f4da8-c2bb-400d-a78f-caf11336d5ec\" (UID: \"944f4da8-c2bb-400d-a78f-caf11336d5ec\") " Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.883622 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.887137 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.889980 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr" (OuterVolumeSpecName: "kube-api-access-mttsr") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "kube-api-access-mttsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.891934 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.892054 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.892273 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.892475 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.896957 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.897270 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.898647 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.901176 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.908805 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.924732 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory" (OuterVolumeSpecName: "inventory") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.942764 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "944f4da8-c2bb-400d-a78f-caf11336d5ec" (UID: "944f4da8-c2bb-400d-a78f-caf11336d5ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974901 4688 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974935 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974946 4688 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974956 4688 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974964 4688 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974974 4688 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974982 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.974993 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975002 4688 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975011 4688 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975021 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mttsr\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-kube-api-access-mttsr\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975030 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975038 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/944f4da8-c2bb-400d-a78f-caf11336d5ec-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:10 crc kubenswrapper[4688]: I1001 16:21:10.975048 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944f4da8-c2bb-400d-a78f-caf11336d5ec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.406260 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" event={"ID":"944f4da8-c2bb-400d-a78f-caf11336d5ec","Type":"ContainerDied","Data":"28debda9718821eff1cd04d127c40cbd8303b213bdd09df8e0c09ab7cb6bcd63"} Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.406299 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28debda9718821eff1cd04d127c40cbd8303b213bdd09df8e0c09ab7cb6bcd63" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.406372 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.519998 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2"] Oct 01 16:21:11 crc kubenswrapper[4688]: E1001 16:21:11.520488 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944f4da8-c2bb-400d-a78f-caf11336d5ec" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.520515 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="944f4da8-c2bb-400d-a78f-caf11336d5ec" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.520763 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="944f4da8-c2bb-400d-a78f-caf11336d5ec" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.521746 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.532251 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2"] Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.534261 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.534774 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.534842 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.534931 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.535106 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.701243 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfzp6\" (UniqueName: \"kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.701445 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.701691 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.701877 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.701995 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.803564 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.803644 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.803691 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.803771 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfzp6\" (UniqueName: \"kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.803831 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.804742 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.809095 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.809544 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.814677 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.829646 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfzp6\" (UniqueName: \"kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-t4nl2\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:11 crc kubenswrapper[4688]: I1001 16:21:11.848143 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:21:12 crc kubenswrapper[4688]: I1001 16:21:12.384445 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2"] Oct 01 16:21:12 crc kubenswrapper[4688]: I1001 16:21:12.418196 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" event={"ID":"da404969-614a-4c71-9ce3-0a15a702aa65","Type":"ContainerStarted","Data":"c37bd29ec985e35c1acf2f20cc96c61de0b22361848d53f90d91df5a891eece1"} Oct 01 16:21:14 crc kubenswrapper[4688]: I1001 16:21:14.450799 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" event={"ID":"da404969-614a-4c71-9ce3-0a15a702aa65","Type":"ContainerStarted","Data":"a2a3eaf085043f2f6e8db17182b2b599a2bf2c019716b3674daccd3e9aea5f96"} Oct 01 16:21:14 crc kubenswrapper[4688]: I1001 16:21:14.469717 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" podStartSLOduration=2.576026437 podStartE2EDuration="3.469698199s" podCreationTimestamp="2025-10-01 16:21:11 +0000 UTC" firstStartedPulling="2025-10-01 16:21:12.386123284 +0000 UTC m=+2061.736763246" lastFinishedPulling="2025-10-01 16:21:13.279795036 +0000 UTC m=+2062.630435008" observedRunningTime="2025-10-01 16:21:14.466483797 +0000 UTC m=+2063.817123759" watchObservedRunningTime="2025-10-01 16:21:14.469698199 +0000 UTC m=+2063.820338161" Oct 01 16:21:41 crc kubenswrapper[4688]: I1001 16:21:41.839853 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:41 crc kubenswrapper[4688]: I1001 16:21:41.842928 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:41 crc kubenswrapper[4688]: I1001 16:21:41.854737 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.015238 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.015302 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.015393 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtlrb\" (UniqueName: \"kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.117827 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.117877 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.117924 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtlrb\" (UniqueName: \"kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.118451 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.118830 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.152871 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtlrb\" (UniqueName: \"kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb\") pod \"redhat-marketplace-sg6bq\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.202098 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.625142 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:42 crc kubenswrapper[4688]: I1001 16:21:42.743489 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerStarted","Data":"acf17b7b7feaf7cf7df8531a113366b931a2ed0c6e8c3268597c73218a4768dc"} Oct 01 16:21:43 crc kubenswrapper[4688]: I1001 16:21:43.754872 4688 generic.go:334] "Generic (PLEG): container finished" podID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerID="e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258" exitCode=0 Oct 01 16:21:43 crc kubenswrapper[4688]: I1001 16:21:43.754945 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerDied","Data":"e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258"} Oct 01 16:21:45 crc kubenswrapper[4688]: I1001 16:21:45.779038 4688 generic.go:334] "Generic (PLEG): container finished" podID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerID="f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e" exitCode=0 Oct 01 16:21:45 crc kubenswrapper[4688]: I1001 16:21:45.779125 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerDied","Data":"f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e"} Oct 01 16:21:46 crc kubenswrapper[4688]: I1001 16:21:46.792941 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerStarted","Data":"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab"} Oct 01 16:21:46 crc kubenswrapper[4688]: I1001 16:21:46.818250 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sg6bq" podStartSLOduration=3.164701845 podStartE2EDuration="5.818230378s" podCreationTimestamp="2025-10-01 16:21:41 +0000 UTC" firstStartedPulling="2025-10-01 16:21:43.757424503 +0000 UTC m=+2093.108064465" lastFinishedPulling="2025-10-01 16:21:46.410953016 +0000 UTC m=+2095.761592998" observedRunningTime="2025-10-01 16:21:46.811677633 +0000 UTC m=+2096.162317615" watchObservedRunningTime="2025-10-01 16:21:46.818230378 +0000 UTC m=+2096.168870340" Oct 01 16:21:52 crc kubenswrapper[4688]: I1001 16:21:52.202515 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:52 crc kubenswrapper[4688]: I1001 16:21:52.203125 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:52 crc kubenswrapper[4688]: I1001 16:21:52.252170 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:52 crc kubenswrapper[4688]: I1001 16:21:52.925883 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:52 crc kubenswrapper[4688]: I1001 16:21:52.986493 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:54 crc kubenswrapper[4688]: I1001 16:21:54.865077 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sg6bq" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="registry-server" containerID="cri-o://69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab" gracePeriod=2 Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.352338 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.491690 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtlrb\" (UniqueName: \"kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb\") pod \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.491762 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities\") pod \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.491910 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content\") pod \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\" (UID: \"ccd52ba6-d315-4217-9af9-8b364edbe2c7\") " Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.492576 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities" (OuterVolumeSpecName: "utilities") pod "ccd52ba6-d315-4217-9af9-8b364edbe2c7" (UID: "ccd52ba6-d315-4217-9af9-8b364edbe2c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.497421 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.497927 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb" (OuterVolumeSpecName: "kube-api-access-qtlrb") pod "ccd52ba6-d315-4217-9af9-8b364edbe2c7" (UID: "ccd52ba6-d315-4217-9af9-8b364edbe2c7"). InnerVolumeSpecName "kube-api-access-qtlrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.504246 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccd52ba6-d315-4217-9af9-8b364edbe2c7" (UID: "ccd52ba6-d315-4217-9af9-8b364edbe2c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.599191 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtlrb\" (UniqueName: \"kubernetes.io/projected/ccd52ba6-d315-4217-9af9-8b364edbe2c7-kube-api-access-qtlrb\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.599539 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd52ba6-d315-4217-9af9-8b364edbe2c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.874266 4688 generic.go:334] "Generic (PLEG): container finished" podID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerID="69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab" exitCode=0 Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.874322 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerDied","Data":"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab"} Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.874356 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sg6bq" event={"ID":"ccd52ba6-d315-4217-9af9-8b364edbe2c7","Type":"ContainerDied","Data":"acf17b7b7feaf7cf7df8531a113366b931a2ed0c6e8c3268597c73218a4768dc"} Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.874377 4688 scope.go:117] "RemoveContainer" containerID="69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.874650 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sg6bq" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.906183 4688 scope.go:117] "RemoveContainer" containerID="f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.919613 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.927103 4688 scope.go:117] "RemoveContainer" containerID="e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.935205 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sg6bq"] Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.974370 4688 scope.go:117] "RemoveContainer" containerID="69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab" Oct 01 16:21:55 crc kubenswrapper[4688]: E1001 16:21:55.974796 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab\": container with ID starting with 69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab not found: ID does not exist" containerID="69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.974869 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab"} err="failed to get container status \"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab\": rpc error: code = NotFound desc = could not find container \"69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab\": container with ID starting with 69a7509946aecf759841c55e97b976a8d3ffed65b2a90613206b616171541dab not found: ID does not exist" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.974904 4688 scope.go:117] "RemoveContainer" containerID="f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e" Oct 01 16:21:55 crc kubenswrapper[4688]: E1001 16:21:55.979761 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e\": container with ID starting with f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e not found: ID does not exist" containerID="f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.979813 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e"} err="failed to get container status \"f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e\": rpc error: code = NotFound desc = could not find container \"f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e\": container with ID starting with f09328ceb7b7fa13b11dfeac5ada908f65ce66e91d88de81a1603d572cb0e65e not found: ID does not exist" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.979840 4688 scope.go:117] "RemoveContainer" containerID="e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258" Oct 01 16:21:55 crc kubenswrapper[4688]: E1001 16:21:55.980157 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258\": container with ID starting with e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258 not found: ID does not exist" containerID="e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258" Oct 01 16:21:55 crc kubenswrapper[4688]: I1001 16:21:55.980181 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258"} err="failed to get container status \"e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258\": rpc error: code = NotFound desc = could not find container \"e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258\": container with ID starting with e4edf8d119f7cdee0c4b437d3aaaea248cc4cf43ccef1c2b4a2f8f37f3d71258 not found: ID does not exist" Oct 01 16:21:57 crc kubenswrapper[4688]: I1001 16:21:57.396207 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" path="/var/lib/kubelet/pods/ccd52ba6-d315-4217-9af9-8b364edbe2c7/volumes" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.915665 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:17 crc kubenswrapper[4688]: E1001 16:22:17.917833 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="registry-server" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.918108 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="registry-server" Oct 01 16:22:17 crc kubenswrapper[4688]: E1001 16:22:17.918213 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="extract-utilities" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.918285 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="extract-utilities" Oct 01 16:22:17 crc kubenswrapper[4688]: E1001 16:22:17.918368 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="extract-content" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.918447 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="extract-content" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.918974 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd52ba6-d315-4217-9af9-8b364edbe2c7" containerName="registry-server" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.921077 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:17 crc kubenswrapper[4688]: I1001 16:22:17.928004 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.038473 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vwbp\" (UniqueName: \"kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.038694 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.038822 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.140579 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.140667 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.140721 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vwbp\" (UniqueName: \"kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.141159 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.141160 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.165353 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vwbp\" (UniqueName: \"kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp\") pod \"certified-operators-jk2v7\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.243337 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:18 crc kubenswrapper[4688]: I1001 16:22:18.865819 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:19 crc kubenswrapper[4688]: I1001 16:22:19.096229 4688 generic.go:334] "Generic (PLEG): container finished" podID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerID="a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756" exitCode=0 Oct 01 16:22:19 crc kubenswrapper[4688]: I1001 16:22:19.096287 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerDied","Data":"a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756"} Oct 01 16:22:19 crc kubenswrapper[4688]: I1001 16:22:19.096325 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerStarted","Data":"e920b4c0199439b6ecb8a1240001f10876fd791fa15dbb0c2013e4dbebaba5ba"} Oct 01 16:22:20 crc kubenswrapper[4688]: I1001 16:22:20.106003 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerStarted","Data":"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1"} Oct 01 16:22:22 crc kubenswrapper[4688]: I1001 16:22:22.134018 4688 generic.go:334] "Generic (PLEG): container finished" podID="da404969-614a-4c71-9ce3-0a15a702aa65" containerID="a2a3eaf085043f2f6e8db17182b2b599a2bf2c019716b3674daccd3e9aea5f96" exitCode=0 Oct 01 16:22:22 crc kubenswrapper[4688]: I1001 16:22:22.134360 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" event={"ID":"da404969-614a-4c71-9ce3-0a15a702aa65","Type":"ContainerDied","Data":"a2a3eaf085043f2f6e8db17182b2b599a2bf2c019716b3674daccd3e9aea5f96"} Oct 01 16:22:22 crc kubenswrapper[4688]: I1001 16:22:22.144114 4688 generic.go:334] "Generic (PLEG): container finished" podID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerID="7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1" exitCode=0 Oct 01 16:22:22 crc kubenswrapper[4688]: I1001 16:22:22.144177 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerDied","Data":"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1"} Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.158129 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerStarted","Data":"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5"} Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.195954 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jk2v7" podStartSLOduration=2.645021099 podStartE2EDuration="6.195936623s" podCreationTimestamp="2025-10-01 16:22:17 +0000 UTC" firstStartedPulling="2025-10-01 16:22:19.097863851 +0000 UTC m=+2128.448503823" lastFinishedPulling="2025-10-01 16:22:22.648779375 +0000 UTC m=+2131.999419347" observedRunningTime="2025-10-01 16:22:23.195229413 +0000 UTC m=+2132.545869385" watchObservedRunningTime="2025-10-01 16:22:23.195936623 +0000 UTC m=+2132.546576595" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.552164 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.663122 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key\") pod \"da404969-614a-4c71-9ce3-0a15a702aa65\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.663189 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0\") pod \"da404969-614a-4c71-9ce3-0a15a702aa65\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.663226 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfzp6\" (UniqueName: \"kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6\") pod \"da404969-614a-4c71-9ce3-0a15a702aa65\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.663372 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle\") pod \"da404969-614a-4c71-9ce3-0a15a702aa65\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.663534 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory\") pod \"da404969-614a-4c71-9ce3-0a15a702aa65\" (UID: \"da404969-614a-4c71-9ce3-0a15a702aa65\") " Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.671813 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6" (OuterVolumeSpecName: "kube-api-access-kfzp6") pod "da404969-614a-4c71-9ce3-0a15a702aa65" (UID: "da404969-614a-4c71-9ce3-0a15a702aa65"). InnerVolumeSpecName "kube-api-access-kfzp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.690786 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "da404969-614a-4c71-9ce3-0a15a702aa65" (UID: "da404969-614a-4c71-9ce3-0a15a702aa65"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.695596 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da404969-614a-4c71-9ce3-0a15a702aa65" (UID: "da404969-614a-4c71-9ce3-0a15a702aa65"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.698309 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "da404969-614a-4c71-9ce3-0a15a702aa65" (UID: "da404969-614a-4c71-9ce3-0a15a702aa65"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.702060 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory" (OuterVolumeSpecName: "inventory") pod "da404969-614a-4c71-9ce3-0a15a702aa65" (UID: "da404969-614a-4c71-9ce3-0a15a702aa65"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.765998 4688 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.766030 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.766039 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da404969-614a-4c71-9ce3-0a15a702aa65-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.766050 4688 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/da404969-614a-4c71-9ce3-0a15a702aa65-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:23 crc kubenswrapper[4688]: I1001 16:22:23.766062 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfzp6\" (UniqueName: \"kubernetes.io/projected/da404969-614a-4c71-9ce3-0a15a702aa65-kube-api-access-kfzp6\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.168760 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" event={"ID":"da404969-614a-4c71-9ce3-0a15a702aa65","Type":"ContainerDied","Data":"c37bd29ec985e35c1acf2f20cc96c61de0b22361848d53f90d91df5a891eece1"} Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.169038 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c37bd29ec985e35c1acf2f20cc96c61de0b22361848d53f90d91df5a891eece1" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.168840 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-t4nl2" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.269408 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz"] Oct 01 16:22:24 crc kubenswrapper[4688]: E1001 16:22:24.270004 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da404969-614a-4c71-9ce3-0a15a702aa65" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.270030 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="da404969-614a-4c71-9ce3-0a15a702aa65" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.270332 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="da404969-614a-4c71-9ce3-0a15a702aa65" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.271276 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.274100 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.274272 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.275173 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.275373 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.275709 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.275880 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.296812 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz"] Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380573 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380626 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380710 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380740 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380801 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.380868 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gqwp\" (UniqueName: \"kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483209 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483276 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483304 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483362 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gqwp\" (UniqueName: \"kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483454 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.483486 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.488107 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.488271 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.488677 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.490230 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.491537 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.508267 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gqwp\" (UniqueName: \"kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:24 crc kubenswrapper[4688]: I1001 16:22:24.588658 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:22:25 crc kubenswrapper[4688]: I1001 16:22:25.052488 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:22:25 crc kubenswrapper[4688]: I1001 16:22:25.052901 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:22:25 crc kubenswrapper[4688]: I1001 16:22:25.179082 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz"] Oct 01 16:22:25 crc kubenswrapper[4688]: W1001 16:22:25.180495 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ddaaf82_b5e6_4960_b79e_77ad2f913b93.slice/crio-4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d WatchSource:0}: Error finding container 4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d: Status 404 returned error can't find the container with id 4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d Oct 01 16:22:26 crc kubenswrapper[4688]: I1001 16:22:26.191843 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" event={"ID":"0ddaaf82-b5e6-4960-b79e-77ad2f913b93","Type":"ContainerStarted","Data":"970d63c761ba0517991acd009d6d739a959b79a46cba4d5bf086b0cf259f9374"} Oct 01 16:22:26 crc kubenswrapper[4688]: I1001 16:22:26.193312 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" event={"ID":"0ddaaf82-b5e6-4960-b79e-77ad2f913b93","Type":"ContainerStarted","Data":"4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d"} Oct 01 16:22:26 crc kubenswrapper[4688]: I1001 16:22:26.211104 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" podStartSLOduration=1.7503130489999998 podStartE2EDuration="2.21108244s" podCreationTimestamp="2025-10-01 16:22:24 +0000 UTC" firstStartedPulling="2025-10-01 16:22:25.18613599 +0000 UTC m=+2134.536775952" lastFinishedPulling="2025-10-01 16:22:25.646905371 +0000 UTC m=+2134.997545343" observedRunningTime="2025-10-01 16:22:26.207130389 +0000 UTC m=+2135.557770361" watchObservedRunningTime="2025-10-01 16:22:26.21108244 +0000 UTC m=+2135.561722402" Oct 01 16:22:28 crc kubenswrapper[4688]: I1001 16:22:28.244039 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:28 crc kubenswrapper[4688]: I1001 16:22:28.244353 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:29 crc kubenswrapper[4688]: I1001 16:22:29.285059 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jk2v7" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="registry-server" probeResult="failure" output=< Oct 01 16:22:29 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:22:29 crc kubenswrapper[4688]: > Oct 01 16:22:38 crc kubenswrapper[4688]: I1001 16:22:38.300571 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:38 crc kubenswrapper[4688]: I1001 16:22:38.376268 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:38 crc kubenswrapper[4688]: I1001 16:22:38.542993 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.306959 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jk2v7" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="registry-server" containerID="cri-o://edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5" gracePeriod=2 Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.790466 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.954926 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities\") pod \"dd50b9e2-37b2-475e-b489-7f71327f8262\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.955742 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities" (OuterVolumeSpecName: "utilities") pod "dd50b9e2-37b2-475e-b489-7f71327f8262" (UID: "dd50b9e2-37b2-475e-b489-7f71327f8262"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.955984 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content\") pod \"dd50b9e2-37b2-475e-b489-7f71327f8262\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.966878 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vwbp\" (UniqueName: \"kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp\") pod \"dd50b9e2-37b2-475e-b489-7f71327f8262\" (UID: \"dd50b9e2-37b2-475e-b489-7f71327f8262\") " Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.967723 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:40 crc kubenswrapper[4688]: I1001 16:22:40.974816 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp" (OuterVolumeSpecName: "kube-api-access-9vwbp") pod "dd50b9e2-37b2-475e-b489-7f71327f8262" (UID: "dd50b9e2-37b2-475e-b489-7f71327f8262"). InnerVolumeSpecName "kube-api-access-9vwbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.000172 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd50b9e2-37b2-475e-b489-7f71327f8262" (UID: "dd50b9e2-37b2-475e-b489-7f71327f8262"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.069819 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vwbp\" (UniqueName: \"kubernetes.io/projected/dd50b9e2-37b2-475e-b489-7f71327f8262-kube-api-access-9vwbp\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.069865 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd50b9e2-37b2-475e-b489-7f71327f8262-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.318646 4688 generic.go:334] "Generic (PLEG): container finished" podID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerID="edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5" exitCode=0 Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.318734 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerDied","Data":"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5"} Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.318802 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk2v7" event={"ID":"dd50b9e2-37b2-475e-b489-7f71327f8262","Type":"ContainerDied","Data":"e920b4c0199439b6ecb8a1240001f10876fd791fa15dbb0c2013e4dbebaba5ba"} Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.318828 4688 scope.go:117] "RemoveContainer" containerID="edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.319700 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk2v7" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.343856 4688 scope.go:117] "RemoveContainer" containerID="7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.374911 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.382204 4688 scope.go:117] "RemoveContainer" containerID="a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.391681 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jk2v7"] Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.416268 4688 scope.go:117] "RemoveContainer" containerID="edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5" Oct 01 16:22:41 crc kubenswrapper[4688]: E1001 16:22:41.416923 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5\": container with ID starting with edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5 not found: ID does not exist" containerID="edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.416963 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5"} err="failed to get container status \"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5\": rpc error: code = NotFound desc = could not find container \"edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5\": container with ID starting with edb45ca5b0d22ad97461e7803eda92ab520c0fd1103deba15153c23f14ace2e5 not found: ID does not exist" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.416991 4688 scope.go:117] "RemoveContainer" containerID="7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1" Oct 01 16:22:41 crc kubenswrapper[4688]: E1001 16:22:41.417278 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1\": container with ID starting with 7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1 not found: ID does not exist" containerID="7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.417305 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1"} err="failed to get container status \"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1\": rpc error: code = NotFound desc = could not find container \"7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1\": container with ID starting with 7745e7affecfa0e89452189d75d8826fd55da722e72e8be7160f6a0236bf8ee1 not found: ID does not exist" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.417323 4688 scope.go:117] "RemoveContainer" containerID="a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756" Oct 01 16:22:41 crc kubenswrapper[4688]: E1001 16:22:41.417605 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756\": container with ID starting with a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756 not found: ID does not exist" containerID="a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756" Oct 01 16:22:41 crc kubenswrapper[4688]: I1001 16:22:41.417639 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756"} err="failed to get container status \"a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756\": rpc error: code = NotFound desc = could not find container \"a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756\": container with ID starting with a37ab4f3889b09a2161d47dc9a194331cdfd6dadf7865ad7589fe6df49ee1756 not found: ID does not exist" Oct 01 16:22:43 crc kubenswrapper[4688]: I1001 16:22:43.398595 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" path="/var/lib/kubelet/pods/dd50b9e2-37b2-475e-b489-7f71327f8262/volumes" Oct 01 16:22:55 crc kubenswrapper[4688]: I1001 16:22:55.053238 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:22:55 crc kubenswrapper[4688]: I1001 16:22:55.054011 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:23:16 crc kubenswrapper[4688]: I1001 16:23:16.634550 4688 generic.go:334] "Generic (PLEG): container finished" podID="0ddaaf82-b5e6-4960-b79e-77ad2f913b93" containerID="970d63c761ba0517991acd009d6d739a959b79a46cba4d5bf086b0cf259f9374" exitCode=0 Oct 01 16:23:16 crc kubenswrapper[4688]: I1001 16:23:16.634670 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" event={"ID":"0ddaaf82-b5e6-4960-b79e-77ad2f913b93","Type":"ContainerDied","Data":"970d63c761ba0517991acd009d6d739a959b79a46cba4d5bf086b0cf259f9374"} Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.041093 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.128777 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.128979 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.129016 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.129132 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.129205 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.129442 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gqwp\" (UniqueName: \"kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp\") pod \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\" (UID: \"0ddaaf82-b5e6-4960-b79e-77ad2f913b93\") " Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.135704 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp" (OuterVolumeSpecName: "kube-api-access-7gqwp") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "kube-api-access-7gqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.136349 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.158475 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory" (OuterVolumeSpecName: "inventory") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.159426 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.164826 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.171819 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0ddaaf82-b5e6-4960-b79e-77ad2f913b93" (UID: "0ddaaf82-b5e6-4960-b79e-77ad2f913b93"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232109 4688 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232153 4688 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232173 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232184 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232196 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gqwp\" (UniqueName: \"kubernetes.io/projected/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-kube-api-access-7gqwp\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.232209 4688 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0ddaaf82-b5e6-4960-b79e-77ad2f913b93-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.650940 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" event={"ID":"0ddaaf82-b5e6-4960-b79e-77ad2f913b93","Type":"ContainerDied","Data":"4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d"} Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.651236 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c316d11029e1814e8774787960f76adc7903751fd4c47871dfea532b97da37d" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.651014 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.754804 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x"] Oct 01 16:23:18 crc kubenswrapper[4688]: E1001 16:23:18.755173 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="extract-utilities" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755188 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="extract-utilities" Oct 01 16:23:18 crc kubenswrapper[4688]: E1001 16:23:18.755212 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="registry-server" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755220 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="registry-server" Oct 01 16:23:18 crc kubenswrapper[4688]: E1001 16:23:18.755244 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddaaf82-b5e6-4960-b79e-77ad2f913b93" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755252 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddaaf82-b5e6-4960-b79e-77ad2f913b93" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:23:18 crc kubenswrapper[4688]: E1001 16:23:18.755272 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="extract-content" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755277 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="extract-content" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755439 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd50b9e2-37b2-475e-b489-7f71327f8262" containerName="registry-server" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.755462 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ddaaf82-b5e6-4960-b79e-77ad2f913b93" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.756040 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.758980 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.759349 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.759477 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.759694 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.759799 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.766681 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x"] Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.841491 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd4k8\" (UniqueName: \"kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.841598 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.841644 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.841665 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.841729 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.943175 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd4k8\" (UniqueName: \"kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.943281 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.943347 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.943384 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.943466 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.947623 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.948291 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.948457 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.954694 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:18 crc kubenswrapper[4688]: I1001 16:23:18.959902 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd4k8\" (UniqueName: \"kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:19 crc kubenswrapper[4688]: I1001 16:23:19.074746 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:23:19 crc kubenswrapper[4688]: I1001 16:23:19.587410 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x"] Oct 01 16:23:19 crc kubenswrapper[4688]: I1001 16:23:19.660196 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" event={"ID":"959f38f8-3df7-422b-9c09-a6fa68d4b8f0","Type":"ContainerStarted","Data":"39bc71e1f2d3638bd5671e0504c956fcf6e11371333939351083486afa1fa4ac"} Oct 01 16:23:20 crc kubenswrapper[4688]: I1001 16:23:20.670693 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" event={"ID":"959f38f8-3df7-422b-9c09-a6fa68d4b8f0","Type":"ContainerStarted","Data":"5aac6b7727d74273534b0c1232bac6e4e21f8d5d4ad73c93fe1a32200724fc11"} Oct 01 16:23:20 crc kubenswrapper[4688]: I1001 16:23:20.697119 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" podStartSLOduration=1.9522962069999998 podStartE2EDuration="2.697101201s" podCreationTimestamp="2025-10-01 16:23:18 +0000 UTC" firstStartedPulling="2025-10-01 16:23:19.595495029 +0000 UTC m=+2188.946134991" lastFinishedPulling="2025-10-01 16:23:20.340300023 +0000 UTC m=+2189.690939985" observedRunningTime="2025-10-01 16:23:20.683627941 +0000 UTC m=+2190.034267913" watchObservedRunningTime="2025-10-01 16:23:20.697101201 +0000 UTC m=+2190.047741163" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.053713 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.054404 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.054458 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.055494 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.055634 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" gracePeriod=600 Oct 01 16:23:25 crc kubenswrapper[4688]: E1001 16:23:25.328930 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.713622 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" exitCode=0 Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.713691 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b"} Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.714073 4688 scope.go:117] "RemoveContainer" containerID="931de179263691b1663c9ab0f98bbeebb63b0503323edea5c88dd374a2e7364f" Oct 01 16:23:25 crc kubenswrapper[4688]: I1001 16:23:25.714658 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:23:25 crc kubenswrapper[4688]: E1001 16:23:25.715095 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:23:36 crc kubenswrapper[4688]: I1001 16:23:36.381631 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:23:36 crc kubenswrapper[4688]: E1001 16:23:36.382712 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:23:51 crc kubenswrapper[4688]: I1001 16:23:51.391003 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:23:51 crc kubenswrapper[4688]: E1001 16:23:51.391579 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:06 crc kubenswrapper[4688]: I1001 16:24:06.381404 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:24:06 crc kubenswrapper[4688]: E1001 16:24:06.382208 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:21 crc kubenswrapper[4688]: I1001 16:24:21.387471 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:24:21 crc kubenswrapper[4688]: E1001 16:24:21.388272 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:32 crc kubenswrapper[4688]: I1001 16:24:32.381396 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:24:32 crc kubenswrapper[4688]: E1001 16:24:32.382494 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:44 crc kubenswrapper[4688]: I1001 16:24:44.380663 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:24:44 crc kubenswrapper[4688]: E1001 16:24:44.381344 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.579010 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.585928 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.613932 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.775558 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.775817 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.775974 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh8z4\" (UniqueName: \"kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.877719 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.877888 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.877978 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh8z4\" (UniqueName: \"kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.878279 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.878392 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.902733 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh8z4\" (UniqueName: \"kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4\") pod \"community-operators-vjr5h\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:56 crc kubenswrapper[4688]: I1001 16:24:56.904305 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:24:57 crc kubenswrapper[4688]: I1001 16:24:57.382243 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:24:57 crc kubenswrapper[4688]: E1001 16:24:57.382887 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:24:57 crc kubenswrapper[4688]: I1001 16:24:57.498245 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:24:57 crc kubenswrapper[4688]: I1001 16:24:57.527255 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerStarted","Data":"588371fa36bd5d6d6ec4e9b0cd67d32342c657630512000800a94b5f7767149d"} Oct 01 16:24:58 crc kubenswrapper[4688]: I1001 16:24:58.536696 4688 generic.go:334] "Generic (PLEG): container finished" podID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerID="6a153dc24a0760fc4ee7fcb4214e5bc0b979a2c85a4bf36972df380565c170a7" exitCode=0 Oct 01 16:24:58 crc kubenswrapper[4688]: I1001 16:24:58.536793 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerDied","Data":"6a153dc24a0760fc4ee7fcb4214e5bc0b979a2c85a4bf36972df380565c170a7"} Oct 01 16:24:58 crc kubenswrapper[4688]: I1001 16:24:58.539443 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:24:59 crc kubenswrapper[4688]: I1001 16:24:59.547799 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerStarted","Data":"e8f25d0018c62a1239f60aeaa7deb554c428373b6a48cf3cb2b1319cd3cfa2f4"} Oct 01 16:25:02 crc kubenswrapper[4688]: I1001 16:25:02.575307 4688 generic.go:334] "Generic (PLEG): container finished" podID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerID="e8f25d0018c62a1239f60aeaa7deb554c428373b6a48cf3cb2b1319cd3cfa2f4" exitCode=0 Oct 01 16:25:02 crc kubenswrapper[4688]: I1001 16:25:02.575398 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerDied","Data":"e8f25d0018c62a1239f60aeaa7deb554c428373b6a48cf3cb2b1319cd3cfa2f4"} Oct 01 16:25:03 crc kubenswrapper[4688]: I1001 16:25:03.586930 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerStarted","Data":"fb5a8d6c36bd3464c7d56e86ea42c4d253aa6bac9610ba126121d6d61505bd02"} Oct 01 16:25:03 crc kubenswrapper[4688]: I1001 16:25:03.611072 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vjr5h" podStartSLOduration=2.995098211 podStartE2EDuration="7.611053686s" podCreationTimestamp="2025-10-01 16:24:56 +0000 UTC" firstStartedPulling="2025-10-01 16:24:58.539214097 +0000 UTC m=+2287.889854059" lastFinishedPulling="2025-10-01 16:25:03.155169552 +0000 UTC m=+2292.505809534" observedRunningTime="2025-10-01 16:25:03.61049205 +0000 UTC m=+2292.961132012" watchObservedRunningTime="2025-10-01 16:25:03.611053686 +0000 UTC m=+2292.961693648" Oct 01 16:25:06 crc kubenswrapper[4688]: I1001 16:25:06.904775 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:06 crc kubenswrapper[4688]: I1001 16:25:06.905097 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:07 crc kubenswrapper[4688]: I1001 16:25:07.949358 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-vjr5h" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="registry-server" probeResult="failure" output=< Oct 01 16:25:07 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:25:07 crc kubenswrapper[4688]: > Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.330354 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qz9z6"] Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.339892 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.361515 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz9z6"] Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.521381 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bwj\" (UniqueName: \"kubernetes.io/projected/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-kube-api-access-m7bwj\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.521717 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-catalog-content\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.521863 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-utilities\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.623724 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7bwj\" (UniqueName: \"kubernetes.io/projected/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-kube-api-access-m7bwj\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.623775 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-catalog-content\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.623809 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-utilities\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.624279 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-catalog-content\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.624298 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-utilities\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.649513 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7bwj\" (UniqueName: \"kubernetes.io/projected/b8cf2c87-9327-4dbd-b969-5a0d4000fd3f-kube-api-access-m7bwj\") pod \"redhat-operators-qz9z6\" (UID: \"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f\") " pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:09 crc kubenswrapper[4688]: I1001 16:25:09.668118 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:10 crc kubenswrapper[4688]: I1001 16:25:10.147057 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz9z6"] Oct 01 16:25:10 crc kubenswrapper[4688]: W1001 16:25:10.150092 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8cf2c87_9327_4dbd_b969_5a0d4000fd3f.slice/crio-507a8e4a5cc6cb2311f90b4fba6a0b7536119812ad08d7e9708784d8bf1b7d06 WatchSource:0}: Error finding container 507a8e4a5cc6cb2311f90b4fba6a0b7536119812ad08d7e9708784d8bf1b7d06: Status 404 returned error can't find the container with id 507a8e4a5cc6cb2311f90b4fba6a0b7536119812ad08d7e9708784d8bf1b7d06 Oct 01 16:25:10 crc kubenswrapper[4688]: I1001 16:25:10.646903 4688 generic.go:334] "Generic (PLEG): container finished" podID="b8cf2c87-9327-4dbd-b969-5a0d4000fd3f" containerID="e8451ed11597afde1bd45d02e2d4a615eee63c42464a50322d60ee8bea3ddb72" exitCode=0 Oct 01 16:25:10 crc kubenswrapper[4688]: I1001 16:25:10.647040 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz9z6" event={"ID":"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f","Type":"ContainerDied","Data":"e8451ed11597afde1bd45d02e2d4a615eee63c42464a50322d60ee8bea3ddb72"} Oct 01 16:25:10 crc kubenswrapper[4688]: I1001 16:25:10.647209 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz9z6" event={"ID":"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f","Type":"ContainerStarted","Data":"507a8e4a5cc6cb2311f90b4fba6a0b7536119812ad08d7e9708784d8bf1b7d06"} Oct 01 16:25:12 crc kubenswrapper[4688]: I1001 16:25:12.382009 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:25:12 crc kubenswrapper[4688]: E1001 16:25:12.382658 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:25:16 crc kubenswrapper[4688]: I1001 16:25:16.954016 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:17 crc kubenswrapper[4688]: I1001 16:25:17.021710 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:17 crc kubenswrapper[4688]: I1001 16:25:17.191888 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:25:18 crc kubenswrapper[4688]: I1001 16:25:18.714651 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vjr5h" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="registry-server" containerID="cri-o://fb5a8d6c36bd3464c7d56e86ea42c4d253aa6bac9610ba126121d6d61505bd02" gracePeriod=2 Oct 01 16:25:19 crc kubenswrapper[4688]: I1001 16:25:19.728439 4688 generic.go:334] "Generic (PLEG): container finished" podID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerID="fb5a8d6c36bd3464c7d56e86ea42c4d253aa6bac9610ba126121d6d61505bd02" exitCode=0 Oct 01 16:25:19 crc kubenswrapper[4688]: I1001 16:25:19.728498 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerDied","Data":"fb5a8d6c36bd3464c7d56e86ea42c4d253aa6bac9610ba126121d6d61505bd02"} Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.082722 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.154188 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh8z4\" (UniqueName: \"kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4\") pod \"6d30cacd-9fc8-4768-ba07-4f80c1332579\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.154661 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities\") pod \"6d30cacd-9fc8-4768-ba07-4f80c1332579\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.154875 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content\") pod \"6d30cacd-9fc8-4768-ba07-4f80c1332579\" (UID: \"6d30cacd-9fc8-4768-ba07-4f80c1332579\") " Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.155236 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities" (OuterVolumeSpecName: "utilities") pod "6d30cacd-9fc8-4768-ba07-4f80c1332579" (UID: "6d30cacd-9fc8-4768-ba07-4f80c1332579"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.155832 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.175083 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4" (OuterVolumeSpecName: "kube-api-access-hh8z4") pod "6d30cacd-9fc8-4768-ba07-4f80c1332579" (UID: "6d30cacd-9fc8-4768-ba07-4f80c1332579"). InnerVolumeSpecName "kube-api-access-hh8z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.195613 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d30cacd-9fc8-4768-ba07-4f80c1332579" (UID: "6d30cacd-9fc8-4768-ba07-4f80c1332579"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.257062 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d30cacd-9fc8-4768-ba07-4f80c1332579-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.257092 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh8z4\" (UniqueName: \"kubernetes.io/projected/6d30cacd-9fc8-4768-ba07-4f80c1332579-kube-api-access-hh8z4\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.747838 4688 generic.go:334] "Generic (PLEG): container finished" podID="b8cf2c87-9327-4dbd-b969-5a0d4000fd3f" containerID="8b5570190a37f6c9e59b444f2aa6c1e9497b75c12f6529c5e747bc91fee1da17" exitCode=0 Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.747918 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz9z6" event={"ID":"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f","Type":"ContainerDied","Data":"8b5570190a37f6c9e59b444f2aa6c1e9497b75c12f6529c5e747bc91fee1da17"} Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.751169 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjr5h" event={"ID":"6d30cacd-9fc8-4768-ba07-4f80c1332579","Type":"ContainerDied","Data":"588371fa36bd5d6d6ec4e9b0cd67d32342c657630512000800a94b5f7767149d"} Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.751199 4688 scope.go:117] "RemoveContainer" containerID="fb5a8d6c36bd3464c7d56e86ea42c4d253aa6bac9610ba126121d6d61505bd02" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.751725 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjr5h" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.779623 4688 scope.go:117] "RemoveContainer" containerID="e8f25d0018c62a1239f60aeaa7deb554c428373b6a48cf3cb2b1319cd3cfa2f4" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.804035 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.811279 4688 scope.go:117] "RemoveContainer" containerID="6a153dc24a0760fc4ee7fcb4214e5bc0b979a2c85a4bf36972df380565c170a7" Oct 01 16:25:21 crc kubenswrapper[4688]: I1001 16:25:21.815793 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vjr5h"] Oct 01 16:25:23 crc kubenswrapper[4688]: I1001 16:25:23.396611 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" path="/var/lib/kubelet/pods/6d30cacd-9fc8-4768-ba07-4f80c1332579/volumes" Oct 01 16:25:23 crc kubenswrapper[4688]: I1001 16:25:23.769651 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz9z6" event={"ID":"b8cf2c87-9327-4dbd-b969-5a0d4000fd3f","Type":"ContainerStarted","Data":"0a5db743c225ead57c6ceb036e625d1f6f46cc04f67a986f1288a53cc0dac8a6"} Oct 01 16:25:23 crc kubenswrapper[4688]: I1001 16:25:23.789601 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qz9z6" podStartSLOduration=2.3289054350000002 podStartE2EDuration="14.789585009s" podCreationTimestamp="2025-10-01 16:25:09 +0000 UTC" firstStartedPulling="2025-10-01 16:25:10.649071533 +0000 UTC m=+2299.999711495" lastFinishedPulling="2025-10-01 16:25:23.109751057 +0000 UTC m=+2312.460391069" observedRunningTime="2025-10-01 16:25:23.786487902 +0000 UTC m=+2313.137127864" watchObservedRunningTime="2025-10-01 16:25:23.789585009 +0000 UTC m=+2313.140224971" Oct 01 16:25:26 crc kubenswrapper[4688]: I1001 16:25:26.382104 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:25:26 crc kubenswrapper[4688]: E1001 16:25:26.383027 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:25:29 crc kubenswrapper[4688]: I1001 16:25:29.668882 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:29 crc kubenswrapper[4688]: I1001 16:25:29.669718 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:29 crc kubenswrapper[4688]: I1001 16:25:29.721807 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:29 crc kubenswrapper[4688]: I1001 16:25:29.873807 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qz9z6" Oct 01 16:25:29 crc kubenswrapper[4688]: I1001 16:25:29.930809 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz9z6"] Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.006056 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.006561 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wq6zm" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="registry-server" containerID="cri-o://134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa" gracePeriod=2 Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.497409 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.640677 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities\") pod \"d15a0ff9-a537-4e8b-8836-5b20538fe681\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.640741 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kstp9\" (UniqueName: \"kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9\") pod \"d15a0ff9-a537-4e8b-8836-5b20538fe681\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.640857 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content\") pod \"d15a0ff9-a537-4e8b-8836-5b20538fe681\" (UID: \"d15a0ff9-a537-4e8b-8836-5b20538fe681\") " Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.641893 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities" (OuterVolumeSpecName: "utilities") pod "d15a0ff9-a537-4e8b-8836-5b20538fe681" (UID: "d15a0ff9-a537-4e8b-8836-5b20538fe681"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.650921 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9" (OuterVolumeSpecName: "kube-api-access-kstp9") pod "d15a0ff9-a537-4e8b-8836-5b20538fe681" (UID: "d15a0ff9-a537-4e8b-8836-5b20538fe681"). InnerVolumeSpecName "kube-api-access-kstp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.742463 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.742736 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kstp9\" (UniqueName: \"kubernetes.io/projected/d15a0ff9-a537-4e8b-8836-5b20538fe681-kube-api-access-kstp9\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.755013 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d15a0ff9-a537-4e8b-8836-5b20538fe681" (UID: "d15a0ff9-a537-4e8b-8836-5b20538fe681"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.840369 4688 generic.go:334] "Generic (PLEG): container finished" podID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerID="134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa" exitCode=0 Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.840432 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerDied","Data":"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa"} Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.840757 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wq6zm" event={"ID":"d15a0ff9-a537-4e8b-8836-5b20538fe681","Type":"ContainerDied","Data":"d61ec1d1903814d8845bdec47079d03a1b948acc4f25f6c50aac80c35312d9ee"} Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.840811 4688 scope.go:117] "RemoveContainer" containerID="134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.840462 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wq6zm" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.844920 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a0ff9-a537-4e8b-8836-5b20538fe681-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.866470 4688 scope.go:117] "RemoveContainer" containerID="15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.877292 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.888441 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wq6zm"] Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.901154 4688 scope.go:117] "RemoveContainer" containerID="9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.937972 4688 scope.go:117] "RemoveContainer" containerID="134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa" Oct 01 16:25:30 crc kubenswrapper[4688]: E1001 16:25:30.938500 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa\": container with ID starting with 134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa not found: ID does not exist" containerID="134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.938586 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa"} err="failed to get container status \"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa\": rpc error: code = NotFound desc = could not find container \"134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa\": container with ID starting with 134ef0a282a456fddbd3059efe16fe1b048255e911d5dd50cb6709ea1624cdfa not found: ID does not exist" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.938612 4688 scope.go:117] "RemoveContainer" containerID="15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f" Oct 01 16:25:30 crc kubenswrapper[4688]: E1001 16:25:30.939188 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f\": container with ID starting with 15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f not found: ID does not exist" containerID="15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.939218 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f"} err="failed to get container status \"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f\": rpc error: code = NotFound desc = could not find container \"15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f\": container with ID starting with 15681289cb4a4ece168053d5fad6bd359bf3c8fbe50fd8cdc4fcc4cdc1b4827f not found: ID does not exist" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.939236 4688 scope.go:117] "RemoveContainer" containerID="9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401" Oct 01 16:25:30 crc kubenswrapper[4688]: E1001 16:25:30.940625 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401\": container with ID starting with 9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401 not found: ID does not exist" containerID="9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401" Oct 01 16:25:30 crc kubenswrapper[4688]: I1001 16:25:30.940685 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401"} err="failed to get container status \"9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401\": rpc error: code = NotFound desc = could not find container \"9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401\": container with ID starting with 9668832fa2888f990ea819e177dcdc05ce78f717fddf77070ba7ebe6adc19401 not found: ID does not exist" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.390009 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" path="/var/lib/kubelet/pods/d15a0ff9-a537-4e8b-8836-5b20538fe681/volumes" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.662134 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd"] Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.662829 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.662912 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.662980 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="extract-content" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663042 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="extract-content" Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.663096 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="extract-utilities" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663145 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="extract-utilities" Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.663204 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="extract-utilities" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663253 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="extract-utilities" Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.663309 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="extract-content" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663361 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="extract-content" Oct 01 16:25:31 crc kubenswrapper[4688]: E1001 16:25:31.663418 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663466 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663699 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d30cacd-9fc8-4768-ba07-4f80c1332579" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.663789 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15a0ff9-a537-4e8b-8836-5b20538fe681" containerName="registry-server" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.665053 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.667101 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.681604 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd"] Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.761144 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.761585 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.761636 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bn6w\" (UniqueName: \"kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.830490 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv"] Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.832416 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.847396 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv"] Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.862925 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863027 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tgk8\" (UniqueName: \"kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863110 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863186 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863223 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bn6w\" (UniqueName: \"kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863257 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863848 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.863955 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.886124 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bn6w\" (UniqueName: \"kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.965061 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.965193 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tgk8\" (UniqueName: \"kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.965224 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.965557 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.965594 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.982991 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:31 crc kubenswrapper[4688]: I1001 16:25:31.983332 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tgk8\" (UniqueName: \"kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.151786 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.344553 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd"] Oct 01 16:25:32 crc kubenswrapper[4688]: W1001 16:25:32.367649 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27aad031_d3e2_46e4_a73c_8a217125a386.slice/crio-e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c WatchSource:0}: Error finding container e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c: Status 404 returned error can't find the container with id e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.715533 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv"] Oct 01 16:25:32 crc kubenswrapper[4688]: W1001 16:25:32.732750 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fa5d90a_0d97_4e6b_a5b7_88030e785f48.slice/crio-c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514 WatchSource:0}: Error finding container c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514: Status 404 returned error can't find the container with id c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514 Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.886079 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" event={"ID":"6fa5d90a-0d97-4e6b-a5b7-88030e785f48","Type":"ContainerStarted","Data":"c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514"} Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.888406 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerStarted","Data":"89890a0925567e4560c86d4764dd1077045b7712d870f3fb495499131f480e81"} Oct 01 16:25:32 crc kubenswrapper[4688]: I1001 16:25:32.888444 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerStarted","Data":"e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c"} Oct 01 16:25:33 crc kubenswrapper[4688]: I1001 16:25:33.899720 4688 generic.go:334] "Generic (PLEG): container finished" podID="27aad031-d3e2-46e4-a73c-8a217125a386" containerID="89890a0925567e4560c86d4764dd1077045b7712d870f3fb495499131f480e81" exitCode=0 Oct 01 16:25:33 crc kubenswrapper[4688]: I1001 16:25:33.899814 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerDied","Data":"89890a0925567e4560c86d4764dd1077045b7712d870f3fb495499131f480e81"} Oct 01 16:25:33 crc kubenswrapper[4688]: I1001 16:25:33.901868 4688 generic.go:334] "Generic (PLEG): container finished" podID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerID="bbcae127ed65783bac5853aaedcb0ca96fb6bce80bf5393f768006001162ac44" exitCode=0 Oct 01 16:25:33 crc kubenswrapper[4688]: I1001 16:25:33.901900 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" event={"ID":"6fa5d90a-0d97-4e6b-a5b7-88030e785f48","Type":"ContainerDied","Data":"bbcae127ed65783bac5853aaedcb0ca96fb6bce80bf5393f768006001162ac44"} Oct 01 16:25:35 crc kubenswrapper[4688]: I1001 16:25:35.924366 4688 generic.go:334] "Generic (PLEG): container finished" podID="27aad031-d3e2-46e4-a73c-8a217125a386" containerID="b4506a0967e3cda2ff4e1035abc86f29a81c0ba95fabc69c43b4fe598e1bc494" exitCode=0 Oct 01 16:25:35 crc kubenswrapper[4688]: I1001 16:25:35.924650 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerDied","Data":"b4506a0967e3cda2ff4e1035abc86f29a81c0ba95fabc69c43b4fe598e1bc494"} Oct 01 16:25:35 crc kubenswrapper[4688]: I1001 16:25:35.928614 4688 generic.go:334] "Generic (PLEG): container finished" podID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerID="620f713e30be607eb0c52e1e4a19a296179d5d53afe4550be10e7f142fc2ea38" exitCode=0 Oct 01 16:25:35 crc kubenswrapper[4688]: I1001 16:25:35.928657 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" event={"ID":"6fa5d90a-0d97-4e6b-a5b7-88030e785f48","Type":"ContainerDied","Data":"620f713e30be607eb0c52e1e4a19a296179d5d53afe4550be10e7f142fc2ea38"} Oct 01 16:25:36 crc kubenswrapper[4688]: I1001 16:25:36.940907 4688 generic.go:334] "Generic (PLEG): container finished" podID="27aad031-d3e2-46e4-a73c-8a217125a386" containerID="953d218b877af9e4c9f73f940272628ba7852c5efe8ae05b4d2d142a3b402c27" exitCode=0 Oct 01 16:25:36 crc kubenswrapper[4688]: I1001 16:25:36.940969 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerDied","Data":"953d218b877af9e4c9f73f940272628ba7852c5efe8ae05b4d2d142a3b402c27"} Oct 01 16:25:36 crc kubenswrapper[4688]: I1001 16:25:36.942577 4688 generic.go:334] "Generic (PLEG): container finished" podID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerID="d57ce9ce699c19564e220ea319448340d7c9e6c4dc97f15b5cb0fc85c7544775" exitCode=0 Oct 01 16:25:36 crc kubenswrapper[4688]: I1001 16:25:36.942603 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" event={"ID":"6fa5d90a-0d97-4e6b-a5b7-88030e785f48","Type":"ContainerDied","Data":"d57ce9ce699c19564e220ea319448340d7c9e6c4dc97f15b5cb0fc85c7544775"} Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.444715 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.459098 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498083 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bn6w\" (UniqueName: \"kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w\") pod \"27aad031-d3e2-46e4-a73c-8a217125a386\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498139 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle\") pod \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498264 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util\") pod \"27aad031-d3e2-46e4-a73c-8a217125a386\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498308 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tgk8\" (UniqueName: \"kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8\") pod \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498349 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle\") pod \"27aad031-d3e2-46e4-a73c-8a217125a386\" (UID: \"27aad031-d3e2-46e4-a73c-8a217125a386\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.498398 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util\") pod \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\" (UID: \"6fa5d90a-0d97-4e6b-a5b7-88030e785f48\") " Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.504200 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle" (OuterVolumeSpecName: "bundle") pod "6fa5d90a-0d97-4e6b-a5b7-88030e785f48" (UID: "6fa5d90a-0d97-4e6b-a5b7-88030e785f48"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.504814 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle" (OuterVolumeSpecName: "bundle") pod "27aad031-d3e2-46e4-a73c-8a217125a386" (UID: "27aad031-d3e2-46e4-a73c-8a217125a386"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.508345 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8" (OuterVolumeSpecName: "kube-api-access-2tgk8") pod "6fa5d90a-0d97-4e6b-a5b7-88030e785f48" (UID: "6fa5d90a-0d97-4e6b-a5b7-88030e785f48"). InnerVolumeSpecName "kube-api-access-2tgk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.508869 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w" (OuterVolumeSpecName: "kube-api-access-2bn6w") pod "27aad031-d3e2-46e4-a73c-8a217125a386" (UID: "27aad031-d3e2-46e4-a73c-8a217125a386"). InnerVolumeSpecName "kube-api-access-2bn6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.518807 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util" (OuterVolumeSpecName: "util") pod "27aad031-d3e2-46e4-a73c-8a217125a386" (UID: "27aad031-d3e2-46e4-a73c-8a217125a386"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.599399 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bn6w\" (UniqueName: \"kubernetes.io/projected/27aad031-d3e2-46e4-a73c-8a217125a386-kube-api-access-2bn6w\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.599427 4688 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.599437 4688 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.599445 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tgk8\" (UniqueName: \"kubernetes.io/projected/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-kube-api-access-2tgk8\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.599453 4688 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27aad031-d3e2-46e4-a73c-8a217125a386-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.647703 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util" (OuterVolumeSpecName: "util") pod "6fa5d90a-0d97-4e6b-a5b7-88030e785f48" (UID: "6fa5d90a-0d97-4e6b-a5b7-88030e785f48"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.700905 4688 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa5d90a-0d97-4e6b-a5b7-88030e785f48-util\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.966399 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" event={"ID":"27aad031-d3e2-46e4-a73c-8a217125a386","Type":"ContainerDied","Data":"e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c"} Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.966727 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e71f4f862d051c0926c3a4b6ea39ec193feaa48fcd0dcf707d080e45069b3e4c" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.966587 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.971184 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" event={"ID":"6fa5d90a-0d97-4e6b-a5b7-88030e785f48","Type":"ContainerDied","Data":"c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514"} Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.971248 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5e639c09c21a0d49bb5efbb7b74595c7b24f128dd2285e0aca9ecd559dc9514" Oct 01 16:25:38 crc kubenswrapper[4688]: I1001 16:25:38.971514 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv" Oct 01 16:25:41 crc kubenswrapper[4688]: I1001 16:25:41.388632 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:25:41 crc kubenswrapper[4688]: E1001 16:25:41.389215 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520074 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb"] Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520760 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="pull" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520773 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="pull" Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520787 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="pull" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520793 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="pull" Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520808 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520815 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520839 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="util" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520845 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="util" Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520865 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="util" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520871 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="util" Oct 01 16:25:43 crc kubenswrapper[4688]: E1001 16:25:43.520880 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.520886 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.521054 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="27aad031-d3e2-46e4-a73c-8a217125a386" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.521070 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa5d90a-0d97-4e6b-a5b7-88030e785f48" containerName="extract" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.521675 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.529800 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb"] Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.592974 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9hst\" (UniqueName: \"kubernetes.io/projected/bd754ddd-7d01-499a-9bfc-448c200a8c93-kube-api-access-z9hst\") pod \"nmstate-operator-858ddd8f98-mv2bb\" (UID: \"bd754ddd-7d01-499a-9bfc-448c200a8c93\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.694688 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9hst\" (UniqueName: \"kubernetes.io/projected/bd754ddd-7d01-499a-9bfc-448c200a8c93-kube-api-access-z9hst\") pod \"nmstate-operator-858ddd8f98-mv2bb\" (UID: \"bd754ddd-7d01-499a-9bfc-448c200a8c93\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.727725 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9hst\" (UniqueName: \"kubernetes.io/projected/bd754ddd-7d01-499a-9bfc-448c200a8c93-kube-api-access-z9hst\") pod \"nmstate-operator-858ddd8f98-mv2bb\" (UID: \"bd754ddd-7d01-499a-9bfc-448c200a8c93\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" Oct 01 16:25:43 crc kubenswrapper[4688]: I1001 16:25:43.840816 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" Oct 01 16:25:44 crc kubenswrapper[4688]: I1001 16:25:44.353053 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb"] Oct 01 16:25:44 crc kubenswrapper[4688]: W1001 16:25:44.358004 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd754ddd_7d01_499a_9bfc_448c200a8c93.slice/crio-c7592557d9714897ed7bb0db1675ef00274c863cb9999699214d64012137cae7 WatchSource:0}: Error finding container c7592557d9714897ed7bb0db1675ef00274c863cb9999699214d64012137cae7: Status 404 returned error can't find the container with id c7592557d9714897ed7bb0db1675ef00274c863cb9999699214d64012137cae7 Oct 01 16:25:45 crc kubenswrapper[4688]: I1001 16:25:45.043792 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" event={"ID":"bd754ddd-7d01-499a-9bfc-448c200a8c93","Type":"ContainerStarted","Data":"c7592557d9714897ed7bb0db1675ef00274c863cb9999699214d64012137cae7"} Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.071515 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" event={"ID":"bd754ddd-7d01-499a-9bfc-448c200a8c93","Type":"ContainerStarted","Data":"8c578c3dcb8c4f68af0d41a99e7da8a4dbc5baa4483d2215c7713fc0c3ef6640"} Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.101221 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-mv2bb" podStartSLOduration=1.781484791 podStartE2EDuration="5.101201013s" podCreationTimestamp="2025-10-01 16:25:43 +0000 UTC" firstStartedPulling="2025-10-01 16:25:44.370479319 +0000 UTC m=+2333.721119281" lastFinishedPulling="2025-10-01 16:25:47.690195541 +0000 UTC m=+2337.040835503" observedRunningTime="2025-10-01 16:25:48.101136981 +0000 UTC m=+2337.451776943" watchObservedRunningTime="2025-10-01 16:25:48.101201013 +0000 UTC m=+2337.451840975" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.140191 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.140381 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" podUID="fb9b4322-784f-439f-8164-e168bb743200" containerName="nmstate-operator" containerID="cri-o://e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154" gracePeriod=30 Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.562367 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.562868 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="nmstate-metrics" containerID="cri-o://43e4d93c5aae71ed0e978c4f32f5f20bc04b9868e2e360f407cd36720955612b" gracePeriod=30 Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.563316 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="kube-rbac-proxy" containerID="cri-o://2dec712b03f821c287b68ee0e39424c67309584f4d4c137e179b064ec083f1ed" gracePeriod=30 Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.613449 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.613670 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" podUID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" containerName="nmstate-webhook" containerID="cri-o://3b90b919bbe18fe24741e5e9fb4f3ee0885a95b894312f2facc174759da8fcd2" gracePeriod=30 Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.654607 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-handler-589l7"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.658709 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-handler-589l7" podUID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" containerName="nmstate-handler" containerID="cri-o://0da36f452c607c8830324d13558bd632b21b2fee48afe5db528270dfe2d95f68" gracePeriod=30 Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.781139 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.783786 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.813276 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk"] Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.884857 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1be1c4ac-558b-4b81-97cf-feb86083dd5b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.885148 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx8dd\" (UniqueName: \"kubernetes.io/projected/1be1c4ac-558b-4b81-97cf-feb86083dd5b-kube-api-access-xx8dd\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.885226 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1be1c4ac-558b-4b81-97cf-feb86083dd5b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.959305 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.988243 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1be1c4ac-558b-4b81-97cf-feb86083dd5b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.988363 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx8dd\" (UniqueName: \"kubernetes.io/projected/1be1c4ac-558b-4b81-97cf-feb86083dd5b-kube-api-access-xx8dd\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.988402 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1be1c4ac-558b-4b81-97cf-feb86083dd5b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:48 crc kubenswrapper[4688]: I1001 16:25:48.993304 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1be1c4ac-558b-4b81-97cf-feb86083dd5b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.009659 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1be1c4ac-558b-4b81-97cf-feb86083dd5b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.022659 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx8dd\" (UniqueName: \"kubernetes.io/projected/1be1c4ac-558b-4b81-97cf-feb86083dd5b-kube-api-access-xx8dd\") pod \"nmstate-console-plugin-6b874cbd85-dwkqk\" (UID: \"1be1c4ac-558b-4b81-97cf-feb86083dd5b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.084053 4688 generic.go:334] "Generic (PLEG): container finished" podID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" containerID="3b90b919bbe18fe24741e5e9fb4f3ee0885a95b894312f2facc174759da8fcd2" exitCode=0 Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.084352 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" event={"ID":"41322bce-e40d-47d7-9a5d-cd4ae91eb2be","Type":"ContainerDied","Data":"3b90b919bbe18fe24741e5e9fb4f3ee0885a95b894312f2facc174759da8fcd2"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.097761 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bqk4\" (UniqueName: \"kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4\") pod \"fb9b4322-784f-439f-8164-e168bb743200\" (UID: \"fb9b4322-784f-439f-8164-e168bb743200\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.108450 4688 generic.go:334] "Generic (PLEG): container finished" podID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" containerID="0da36f452c607c8830324d13558bd632b21b2fee48afe5db528270dfe2d95f68" exitCode=0 Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.108748 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-589l7" event={"ID":"33578dc7-7ec3-4972-a650-2202a7f7bfa1","Type":"ContainerDied","Data":"0da36f452c607c8830324d13558bd632b21b2fee48afe5db528270dfe2d95f68"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.108868 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-589l7" event={"ID":"33578dc7-7ec3-4972-a650-2202a7f7bfa1","Type":"ContainerDied","Data":"58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.108937 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b6375354e37db13c53682167a0e70b3138b1cb3284bcaf6e558a8c8fb4ca22" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.113993 4688 generic.go:334] "Generic (PLEG): container finished" podID="fb9b4322-784f-439f-8164-e168bb743200" containerID="e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154" exitCode=0 Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.114088 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" event={"ID":"fb9b4322-784f-439f-8164-e168bb743200","Type":"ContainerDied","Data":"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.114126 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" event={"ID":"fb9b4322-784f-439f-8164-e168bb743200","Type":"ContainerDied","Data":"01675c9d6bac212f45dea06757ac20a65b1ec10f61acbea2c385e95efde8963c"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.114148 4688 scope.go:117] "RemoveContainer" containerID="e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.114345 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.140567 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4" (OuterVolumeSpecName: "kube-api-access-4bqk4") pod "fb9b4322-784f-439f-8164-e168bb743200" (UID: "fb9b4322-784f-439f-8164-e168bb743200"). InnerVolumeSpecName "kube-api-access-4bqk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.143299 4688 generic.go:334] "Generic (PLEG): container finished" podID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerID="2dec712b03f821c287b68ee0e39424c67309584f4d4c137e179b064ec083f1ed" exitCode=0 Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.143332 4688 generic.go:334] "Generic (PLEG): container finished" podID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerID="43e4d93c5aae71ed0e978c4f32f5f20bc04b9868e2e360f407cd36720955612b" exitCode=0 Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.144802 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerDied","Data":"2dec712b03f821c287b68ee0e39424c67309584f4d4c137e179b064ec083f1ed"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.144916 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerDied","Data":"43e4d93c5aae71ed0e978c4f32f5f20bc04b9868e2e360f407cd36720955612b"} Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.156476 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.214780 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bqk4\" (UniqueName: \"kubernetes.io/projected/fb9b4322-784f-439f-8164-e168bb743200-kube-api-access-4bqk4\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.224083 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-mjqzm"] Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.224479 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" containerName="nmstate-handler" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.224495 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" containerName="nmstate-handler" Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.224536 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9b4322-784f-439f-8164-e168bb743200" containerName="nmstate-operator" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.224544 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9b4322-784f-439f-8164-e168bb743200" containerName="nmstate-operator" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.224755 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9b4322-784f-439f-8164-e168bb743200" containerName="nmstate-operator" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.224771 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" containerName="nmstate-handler" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.225378 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.226610 4688 scope.go:117] "RemoveContainer" containerID="e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154" Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.226995 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154\": container with ID starting with e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154 not found: ID does not exist" containerID="e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.227016 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154"} err="failed to get container status \"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154\": rpc error: code = NotFound desc = could not find container \"e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154\": container with ID starting with e031197ac9f8a5e0dd9594d001519b54110c44f37ec113f2bd18d508fc539154 not found: ID does not exist" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.276762 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.316017 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock\") pod \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.316055 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket\") pod \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.316124 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket\") pod \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.316267 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxrzv\" (UniqueName: \"kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv\") pod \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\" (UID: \"33578dc7-7ec3-4972-a650-2202a7f7bfa1\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.317511 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock" (OuterVolumeSpecName: "nmstate-lock") pod "33578dc7-7ec3-4972-a650-2202a7f7bfa1" (UID: "33578dc7-7ec3-4972-a650-2202a7f7bfa1"). InnerVolumeSpecName "nmstate-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.317552 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket" (OuterVolumeSpecName: "dbus-socket") pod "33578dc7-7ec3-4972-a650-2202a7f7bfa1" (UID: "33578dc7-7ec3-4972-a650-2202a7f7bfa1"). InnerVolumeSpecName "dbus-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.317567 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket" (OuterVolumeSpecName: "ovs-socket") pod "33578dc7-7ec3-4972-a650-2202a7f7bfa1" (UID: "33578dc7-7ec3-4972-a650-2202a7f7bfa1"). InnerVolumeSpecName "ovs-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.324545 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv" (OuterVolumeSpecName: "kube-api-access-sxrzv") pod "33578dc7-7ec3-4972-a650-2202a7f7bfa1" (UID: "33578dc7-7ec3-4972-a650-2202a7f7bfa1"). InnerVolumeSpecName "kube-api-access-sxrzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.418892 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-ovs-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419028 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-nmstate-lock\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419058 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-dbus-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419217 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw47p\" (UniqueName: \"kubernetes.io/projected/7e9edca6-ee2b-4366-be0a-7c974e4865d7-kube-api-access-kw47p\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419299 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxrzv\" (UniqueName: \"kubernetes.io/projected/33578dc7-7ec3-4972-a650-2202a7f7bfa1-kube-api-access-sxrzv\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419314 4688 reconciler_common.go:293] "Volume detached for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-nmstate-lock\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419327 4688 reconciler_common.go:293] "Volume detached for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-dbus-socket\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.419553 4688 reconciler_common.go:293] "Volume detached for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33578dc7-7ec3-4972-a650-2202a7f7bfa1-ovs-socket\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.462967 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.475312 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.495619 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hnffz"] Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.523968 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw47p\" (UniqueName: \"kubernetes.io/projected/7e9edca6-ee2b-4366-be0a-7c974e4865d7-kube-api-access-kw47p\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.524030 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-ovs-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.524139 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-nmstate-lock\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.524168 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-dbus-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.524590 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-dbus-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.524947 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-ovs-socket\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.525116 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7e9edca6-ee2b-4366-be0a-7c974e4865d7-nmstate-lock\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.592036 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw47p\" (UniqueName: \"kubernetes.io/projected/7e9edca6-ee2b-4366-be0a-7c974e4865d7-kube-api-access-kw47p\") pod \"nmstate-handler-mjqzm\" (UID: \"7e9edca6-ee2b-4366-be0a-7c974e4865d7\") " pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.617492 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk"] Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.617920 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" containerName="nmstate-webhook" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.617938 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" containerName="nmstate-webhook" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.618164 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" containerName="nmstate-webhook" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.618820 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.625166 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair\") pod \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.625324 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4zwr\" (UniqueName: \"kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr\") pod \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\" (UID: \"41322bce-e40d-47d7-9a5d-cd4ae91eb2be\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.641983 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk"] Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.669499 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr" (OuterVolumeSpecName: "kube-api-access-n4zwr") pod "41322bce-e40d-47d7-9a5d-cd4ae91eb2be" (UID: "41322bce-e40d-47d7-9a5d-cd4ae91eb2be"). InnerVolumeSpecName "kube-api-access-n4zwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.679605 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair" (OuterVolumeSpecName: "tls-key-pair") pod "41322bce-e40d-47d7-9a5d-cd4ae91eb2be" (UID: "41322bce-e40d-47d7-9a5d-cd4ae91eb2be"). InnerVolumeSpecName "tls-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.729207 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccdzw\" (UniqueName: \"kubernetes.io/projected/67f45b7a-07fb-4cf3-b63d-832139f2bccd-kube-api-access-ccdzw\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.729317 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/67f45b7a-07fb-4cf3-b63d-832139f2bccd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.729398 4688 reconciler_common.go:293] "Volume detached for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-tls-key-pair\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.729415 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4zwr\" (UniqueName: \"kubernetes.io/projected/41322bce-e40d-47d7-9a5d-cd4ae91eb2be-kube-api-access-n4zwr\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.737716 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.815504 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk"] Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.834061 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qdqd\" (UniqueName: \"kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd\") pod \"b03e32c5-d2eb-4077-a598-06ad31d0284b\" (UID: \"b03e32c5-d2eb-4077-a598-06ad31d0284b\") " Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.834920 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/67f45b7a-07fb-4cf3-b63d-832139f2bccd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.835019 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccdzw\" (UniqueName: \"kubernetes.io/projected/67f45b7a-07fb-4cf3-b63d-832139f2bccd-kube-api-access-ccdzw\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.840695 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/67f45b7a-07fb-4cf3-b63d-832139f2bccd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.851473 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd" (OuterVolumeSpecName: "kube-api-access-6qdqd") pod "b03e32c5-d2eb-4077-a598-06ad31d0284b" (UID: "b03e32c5-d2eb-4077-a598-06ad31d0284b"). InnerVolumeSpecName "kube-api-access-6qdqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.867291 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.885543 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccdzw\" (UniqueName: \"kubernetes.io/projected/67f45b7a-07fb-4cf3-b63d-832139f2bccd-kube-api-access-ccdzw\") pod \"nmstate-webhook-6cdbc54649-x9vxk\" (UID: \"67f45b7a-07fb-4cf3-b63d-832139f2bccd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.937060 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qdqd\" (UniqueName: \"kubernetes.io/projected/b03e32c5-d2eb-4077-a598-06ad31d0284b-kube-api-access-6qdqd\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.971378 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.985572 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh"] Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.986047 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="kube-rbac-proxy" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.986064 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="kube-rbac-proxy" Oct 01 16:25:49 crc kubenswrapper[4688]: E1001 16:25:49.986083 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="nmstate-metrics" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.986089 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="nmstate-metrics" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.986253 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="kube-rbac-proxy" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.986278 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" containerName="nmstate-metrics" Oct 01 16:25:49 crc kubenswrapper[4688]: I1001 16:25:49.987511 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.144666 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x26w6\" (UniqueName: \"kubernetes.io/projected/8e78e6a0-4428-40b7-b819-6d798cf9983b-kube-api-access-x26w6\") pod \"nmstate-metrics-fdff9cb8d-2x5vh\" (UID: \"8e78e6a0-4428-40b7-b819-6d798cf9983b\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.160648 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.190021 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" event={"ID":"b03e32c5-d2eb-4077-a598-06ad31d0284b","Type":"ContainerDied","Data":"4f464b63dc720eb74068ebc7c08f668cdb3d0d6a224a2f4159fcf7684e883805"} Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.190068 4688 scope.go:117] "RemoveContainer" containerID="2dec712b03f821c287b68ee0e39424c67309584f4d4c137e179b064ec083f1ed" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.190163 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-fj6km" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.215224 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" event={"ID":"1be1c4ac-558b-4b81-97cf-feb86083dd5b","Type":"ContainerStarted","Data":"0ec16fa6ec7fa607825ac0e0b8579510a7d7e6efd7717b8f3b386aff41b45aa7"} Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.246714 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x26w6\" (UniqueName: \"kubernetes.io/projected/8e78e6a0-4428-40b7-b819-6d798cf9983b-kube-api-access-x26w6\") pod \"nmstate-metrics-fdff9cb8d-2x5vh\" (UID: \"8e78e6a0-4428-40b7-b819-6d798cf9983b\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.258722 4688 scope.go:117] "RemoveContainer" containerID="43e4d93c5aae71ed0e978c4f32f5f20bc04b9868e2e360f407cd36720955612b" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.265429 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.265507 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-gzngc" event={"ID":"41322bce-e40d-47d7-9a5d-cd4ae91eb2be","Type":"ContainerDied","Data":"4df989b61c2dfdd963caf47df95de0eba7d2f3c6ddc553a9028bb2d883560f6f"} Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.272492 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x26w6\" (UniqueName: \"kubernetes.io/projected/8e78e6a0-4428-40b7-b819-6d798cf9983b-kube-api-access-x26w6\") pod \"nmstate-metrics-fdff9cb8d-2x5vh\" (UID: \"8e78e6a0-4428-40b7-b819-6d798cf9983b\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.272584 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.283684 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-589l7" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.283815 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mjqzm" event={"ID":"7e9edca6-ee2b-4366-be0a-7c974e4865d7","Type":"ContainerStarted","Data":"fa39328651c6d62460b02a4e61c65bd8ca86b5b0957f1b0225b3685f7c600b02"} Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.306066 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-fj6km"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.346331 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-handler-589l7"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.353515 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.384490 4688 scope.go:117] "RemoveContainer" containerID="3b90b919bbe18fe24741e5e9fb4f3ee0885a95b894312f2facc174759da8fcd2" Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.389943 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-handler-589l7"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.415749 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.432509 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-gzngc"] Oct 01 16:25:50 crc kubenswrapper[4688]: I1001 16:25:50.810107 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk"] Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.155940 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh"] Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.310185 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" event={"ID":"67f45b7a-07fb-4cf3-b63d-832139f2bccd","Type":"ContainerStarted","Data":"e53a985e9e85539964549a1c1b7d88e598c76a13af32afd2e8301d22dba073b4"} Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.323983 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" event={"ID":"8e78e6a0-4428-40b7-b819-6d798cf9983b","Type":"ContainerStarted","Data":"d583f5d4938ff8e98a1c31c86257b73b2ff48ffe0eebaea43a4b0c942e5a4bfe"} Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.409940 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33578dc7-7ec3-4972-a650-2202a7f7bfa1" path="/var/lib/kubelet/pods/33578dc7-7ec3-4972-a650-2202a7f7bfa1/volumes" Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.422444 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41322bce-e40d-47d7-9a5d-cd4ae91eb2be" path="/var/lib/kubelet/pods/41322bce-e40d-47d7-9a5d-cd4ae91eb2be/volumes" Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.424441 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03e32c5-d2eb-4077-a598-06ad31d0284b" path="/var/lib/kubelet/pods/b03e32c5-d2eb-4077-a598-06ad31d0284b/volumes" Oct 01 16:25:51 crc kubenswrapper[4688]: I1001 16:25:51.432669 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb9b4322-784f-439f-8164-e168bb743200" path="/var/lib/kubelet/pods/fb9b4322-784f-439f-8164-e168bb743200/volumes" Oct 01 16:25:54 crc kubenswrapper[4688]: I1001 16:25:54.380659 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:25:54 crc kubenswrapper[4688]: E1001 16:25:54.381420 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.407795 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" event={"ID":"8e78e6a0-4428-40b7-b819-6d798cf9983b","Type":"ContainerStarted","Data":"14f1419f24886dc763b9938b1b3902bffb73905ac2615ff18935cfe5eaf80609"} Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.408147 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" event={"ID":"67f45b7a-07fb-4cf3-b63d-832139f2bccd","Type":"ContainerStarted","Data":"23455858041f27b58e8b9b9c359ba232a78198ddfb548adca0c3feb2b1033af1"} Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.408162 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" event={"ID":"1be1c4ac-558b-4b81-97cf-feb86083dd5b","Type":"ContainerStarted","Data":"b6ded06430af1c4f808d18aa6edf009dbc3a7053630d0822da83f87f48501cf3"} Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.408177 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.426279 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dwkqk" podStartSLOduration=2.646626515 podStartE2EDuration="7.426261872s" podCreationTimestamp="2025-10-01 16:25:48 +0000 UTC" firstStartedPulling="2025-10-01 16:25:49.81332979 +0000 UTC m=+2339.163969752" lastFinishedPulling="2025-10-01 16:25:54.592965147 +0000 UTC m=+2343.943605109" observedRunningTime="2025-10-01 16:25:55.414664191 +0000 UTC m=+2344.765304153" watchObservedRunningTime="2025-10-01 16:25:55.426261872 +0000 UTC m=+2344.776901834" Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.443861 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" podStartSLOduration=2.679253223 podStartE2EDuration="6.443832043s" podCreationTimestamp="2025-10-01 16:25:49 +0000 UTC" firstStartedPulling="2025-10-01 16:25:50.832365901 +0000 UTC m=+2340.183005863" lastFinishedPulling="2025-10-01 16:25:54.596944721 +0000 UTC m=+2343.947584683" observedRunningTime="2025-10-01 16:25:55.440175729 +0000 UTC m=+2344.790815711" watchObservedRunningTime="2025-10-01 16:25:55.443832043 +0000 UTC m=+2344.794472005" Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.470553 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.470757 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" podUID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" containerName="nmstate-console-plugin" containerID="cri-o://ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5" gracePeriod=30 Oct 01 16:25:55 crc kubenswrapper[4688]: I1001 16:25:55.981360 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.069052 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") pod \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.069129 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v65kn\" (UniqueName: \"kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn\") pod \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.069281 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf\") pod \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\" (UID: \"2a77c482-2518-4f89-bcd1-1b9c4e616d30\") " Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.075039 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert" (OuterVolumeSpecName: "plugin-serving-cert") pod "2a77c482-2518-4f89-bcd1-1b9c4e616d30" (UID: "2a77c482-2518-4f89-bcd1-1b9c4e616d30"). InnerVolumeSpecName "plugin-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.088721 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn" (OuterVolumeSpecName: "kube-api-access-v65kn") pod "2a77c482-2518-4f89-bcd1-1b9c4e616d30" (UID: "2a77c482-2518-4f89-bcd1-1b9c4e616d30"). InnerVolumeSpecName "kube-api-access-v65kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.105321 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf" (OuterVolumeSpecName: "nginx-conf") pod "2a77c482-2518-4f89-bcd1-1b9c4e616d30" (UID: "2a77c482-2518-4f89-bcd1-1b9c4e616d30"). InnerVolumeSpecName "nginx-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.171990 4688 reconciler_common.go:293] "Volume detached for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2a77c482-2518-4f89-bcd1-1b9c4e616d30-nginx-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.172021 4688 reconciler_common.go:293] "Volume detached for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a77c482-2518-4f89-bcd1-1b9c4e616d30-plugin-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.172031 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v65kn\" (UniqueName: \"kubernetes.io/projected/2a77c482-2518-4f89-bcd1-1b9c4e616d30-kube-api-access-v65kn\") on node \"crc\" DevicePath \"\"" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.410433 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mjqzm" event={"ID":"7e9edca6-ee2b-4366-be0a-7c974e4865d7","Type":"ContainerStarted","Data":"7a96441a3ba98e7accc533bcdf7d87f51ff9fc44c7c9bd57e3e6cc3a44090831"} Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.410742 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.413243 4688 generic.go:334] "Generic (PLEG): container finished" podID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" containerID="ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5" exitCode=0 Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.413289 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" event={"ID":"2a77c482-2518-4f89-bcd1-1b9c4e616d30","Type":"ContainerDied","Data":"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5"} Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.413341 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" event={"ID":"2a77c482-2518-4f89-bcd1-1b9c4e616d30","Type":"ContainerDied","Data":"365f8a944ed52630f9dd84d5e2b35716bbe3342df83682be99f30a77498fd07c"} Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.413360 4688 scope.go:117] "RemoveContainer" containerID="ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.413310 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.431865 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-mjqzm" podStartSLOduration=2.8598719360000002 podStartE2EDuration="7.43184763s" podCreationTimestamp="2025-10-01 16:25:49 +0000 UTC" firstStartedPulling="2025-10-01 16:25:50.021722634 +0000 UTC m=+2339.372362596" lastFinishedPulling="2025-10-01 16:25:54.593698338 +0000 UTC m=+2343.944338290" observedRunningTime="2025-10-01 16:25:56.429485542 +0000 UTC m=+2345.780125504" watchObservedRunningTime="2025-10-01 16:25:56.43184763 +0000 UTC m=+2345.782487592" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.468161 4688 scope.go:117] "RemoveContainer" containerID="ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5" Oct 01 16:25:56 crc kubenswrapper[4688]: E1001 16:25:56.471870 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5\": container with ID starting with ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5 not found: ID does not exist" containerID="ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.471946 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5"} err="failed to get container status \"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5\": rpc error: code = NotFound desc = could not find container \"ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5\": container with ID starting with ff7a539687dba7e9cff059ad8871106f5a8786c2ed163d910a49bb1ea3f582c5 not found: ID does not exist" Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.479785 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 16:25:56 crc kubenswrapper[4688]: I1001 16:25:56.501636 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-zpjcz"] Oct 01 16:25:57 crc kubenswrapper[4688]: I1001 16:25:57.396258 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" path="/var/lib/kubelet/pods/2a77c482-2518-4f89-bcd1-1b9c4e616d30/volumes" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.325450 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:25:59 crc kubenswrapper[4688]: E1001 16:25:59.326235 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" containerName="nmstate-console-plugin" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.326248 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" containerName="nmstate-console-plugin" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.326425 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a77c482-2518-4f89-bcd1-1b9c4e616d30" containerName="nmstate-console-plugin" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.331556 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.346350 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.451787 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.451962 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.452019 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzvmn\" (UniqueName: \"kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.517095 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" event={"ID":"8e78e6a0-4428-40b7-b819-6d798cf9983b","Type":"ContainerStarted","Data":"12b9f101aac9d3af5f1cde9f6603eb3cf0d43489b537ee53850ae3d8cf80d62f"} Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.556155 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.556233 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzvmn\" (UniqueName: \"kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.556476 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.579686 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.583354 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.583440 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-2x5vh" podStartSLOduration=3.163022339 podStartE2EDuration="10.583414227s" podCreationTimestamp="2025-10-01 16:25:49 +0000 UTC" firstStartedPulling="2025-10-01 16:25:51.199750908 +0000 UTC m=+2340.550390870" lastFinishedPulling="2025-10-01 16:25:58.620142796 +0000 UTC m=+2347.970782758" observedRunningTime="2025-10-01 16:25:59.548645405 +0000 UTC m=+2348.899285367" watchObservedRunningTime="2025-10-01 16:25:59.583414227 +0000 UTC m=+2348.934054189" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.620802 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzvmn\" (UniqueName: \"kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn\") pod \"metallb-operator-controller-manager-6498d9c794-fhhrl\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.669182 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.792279 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-756f8755cd-ml958"] Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.799602 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.802914 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-756f8755cd-ml958"] Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.881863 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5z2c\" (UniqueName: \"kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.881921 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.882012 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.984078 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.984214 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:25:59 crc kubenswrapper[4688]: I1001 16:25:59.984324 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5z2c\" (UniqueName: \"kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:00 crc kubenswrapper[4688]: I1001 16:26:00.002111 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:00 crc kubenswrapper[4688]: I1001 16:26:00.022372 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5z2c\" (UniqueName: \"kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:00 crc kubenswrapper[4688]: I1001 16:26:00.042263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert\") pod \"metallb-operator-webhook-server-756f8755cd-ml958\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:00 crc kubenswrapper[4688]: I1001 16:26:00.131932 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:00 crc kubenswrapper[4688]: I1001 16:26:00.562192 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:26:00 crc kubenswrapper[4688]: W1001 16:26:00.579820 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod478d5733_100a_438f_84f6_ecd0a041f7ce.slice/crio-b984232ec35d4c044864753d371b8faac94a622a831e195ded08d79b126426bb WatchSource:0}: Error finding container b984232ec35d4c044864753d371b8faac94a622a831e195ded08d79b126426bb: Status 404 returned error can't find the container with id b984232ec35d4c044864753d371b8faac94a622a831e195ded08d79b126426bb Oct 01 16:26:01 crc kubenswrapper[4688]: I1001 16:26:01.175231 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-756f8755cd-ml958"] Oct 01 16:26:01 crc kubenswrapper[4688]: W1001 16:26:01.178712 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23ec7644_389e_4fb6_977a_a1a7c69b3d7e.slice/crio-bd022b254f958ba1236eac612f5032ed4f9f07515bce6bdf9389c76d3a481848 WatchSource:0}: Error finding container bd022b254f958ba1236eac612f5032ed4f9f07515bce6bdf9389c76d3a481848: Status 404 returned error can't find the container with id bd022b254f958ba1236eac612f5032ed4f9f07515bce6bdf9389c76d3a481848 Oct 01 16:26:01 crc kubenswrapper[4688]: I1001 16:26:01.542701 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" event={"ID":"478d5733-100a-438f-84f6-ecd0a041f7ce","Type":"ContainerStarted","Data":"b984232ec35d4c044864753d371b8faac94a622a831e195ded08d79b126426bb"} Oct 01 16:26:01 crc kubenswrapper[4688]: I1001 16:26:01.543665 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" event={"ID":"23ec7644-389e-4fb6-977a-a1a7c69b3d7e","Type":"ContainerStarted","Data":"bd022b254f958ba1236eac612f5032ed4f9f07515bce6bdf9389c76d3a481848"} Oct 01 16:26:01 crc kubenswrapper[4688]: I1001 16:26:01.692679 4688 scope.go:117] "RemoveContainer" containerID="0da36f452c607c8830324d13558bd632b21b2fee48afe5db528270dfe2d95f68" Oct 01 16:26:04 crc kubenswrapper[4688]: I1001 16:26:04.948158 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-mjqzm" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.381085 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:26:09 crc kubenswrapper[4688]: E1001 16:26:09.381905 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.638368 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" event={"ID":"23ec7644-389e-4fb6-977a-a1a7c69b3d7e","Type":"ContainerStarted","Data":"30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8"} Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.639283 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.641230 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" event={"ID":"478d5733-100a-438f-84f6-ecd0a041f7ce","Type":"ContainerStarted","Data":"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323"} Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.641889 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.665057 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" podStartSLOduration=2.690291353 podStartE2EDuration="10.665034188s" podCreationTimestamp="2025-10-01 16:25:59 +0000 UTC" firstStartedPulling="2025-10-01 16:26:01.181727919 +0000 UTC m=+2350.532367871" lastFinishedPulling="2025-10-01 16:26:09.156470734 +0000 UTC m=+2358.507110706" observedRunningTime="2025-10-01 16:26:09.656457613 +0000 UTC m=+2359.007097585" watchObservedRunningTime="2025-10-01 16:26:09.665034188 +0000 UTC m=+2359.015674150" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.690777 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" podStartSLOduration=2.148268735 podStartE2EDuration="10.690735161s" podCreationTimestamp="2025-10-01 16:25:59 +0000 UTC" firstStartedPulling="2025-10-01 16:26:00.588231593 +0000 UTC m=+2349.938871555" lastFinishedPulling="2025-10-01 16:26:09.130698019 +0000 UTC m=+2358.481337981" observedRunningTime="2025-10-01 16:26:09.689921478 +0000 UTC m=+2359.040561460" watchObservedRunningTime="2025-10-01 16:26:09.690735161 +0000 UTC m=+2359.041375113" Oct 01 16:26:09 crc kubenswrapper[4688]: I1001 16:26:09.979802 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-x9vxk" Oct 01 16:26:20 crc kubenswrapper[4688]: I1001 16:26:20.179745 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:26:20 crc kubenswrapper[4688]: I1001 16:26:20.288310 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf"] Oct 01 16:26:20 crc kubenswrapper[4688]: I1001 16:26:20.288755 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" podUID="53e2aa94-6e98-416b-b112-042f4444755d" containerName="webhook-server" containerID="cri-o://ece8ec933b7323a08ea17ba7f71f3a3c261c3cd874f7001a1bc8270670ea5f6d" gracePeriod=2 Oct 01 16:26:20 crc kubenswrapper[4688]: I1001 16:26:20.321248 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf"] Oct 01 16:26:20 crc kubenswrapper[4688]: I1001 16:26:20.771848 4688 generic.go:334] "Generic (PLEG): container finished" podID="53e2aa94-6e98-416b-b112-042f4444755d" containerID="ece8ec933b7323a08ea17ba7f71f3a3c261c3cd874f7001a1bc8270670ea5f6d" exitCode=0 Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.317026 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.402379 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvczd\" (UniqueName: \"kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd\") pod \"53e2aa94-6e98-416b-b112-042f4444755d\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.402420 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert\") pod \"53e2aa94-6e98-416b-b112-042f4444755d\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.402444 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert\") pod \"53e2aa94-6e98-416b-b112-042f4444755d\" (UID: \"53e2aa94-6e98-416b-b112-042f4444755d\") " Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.413879 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "53e2aa94-6e98-416b-b112-042f4444755d" (UID: "53e2aa94-6e98-416b-b112-042f4444755d"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.426243 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd" (OuterVolumeSpecName: "kube-api-access-kvczd") pod "53e2aa94-6e98-416b-b112-042f4444755d" (UID: "53e2aa94-6e98-416b-b112-042f4444755d"). InnerVolumeSpecName "kube-api-access-kvczd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.430791 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "53e2aa94-6e98-416b-b112-042f4444755d" (UID: "53e2aa94-6e98-416b-b112-042f4444755d"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.505079 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvczd\" (UniqueName: \"kubernetes.io/projected/53e2aa94-6e98-416b-b112-042f4444755d-kube-api-access-kvczd\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.505116 4688 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.505130 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/53e2aa94-6e98-416b-b112-042f4444755d-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.781717 4688 scope.go:117] "RemoveContainer" containerID="ece8ec933b7323a08ea17ba7f71f3a3c261c3cd874f7001a1bc8270670ea5f6d" Oct 01 16:26:21 crc kubenswrapper[4688]: I1001 16:26:21.781749 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57dcd5bbd8-lnttf" Oct 01 16:26:23 crc kubenswrapper[4688]: I1001 16:26:23.401869 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53e2aa94-6e98-416b-b112-042f4444755d" path="/var/lib/kubelet/pods/53e2aa94-6e98-416b-b112-042f4444755d/volumes" Oct 01 16:26:24 crc kubenswrapper[4688]: I1001 16:26:24.381841 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:26:24 crc kubenswrapper[4688]: E1001 16:26:24.382080 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:26:36 crc kubenswrapper[4688]: I1001 16:26:36.381134 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:26:36 crc kubenswrapper[4688]: E1001 16:26:36.381837 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:26:39 crc kubenswrapper[4688]: I1001 16:26:39.671482 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:26:39 crc kubenswrapper[4688]: I1001 16:26:39.755088 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 16:26:39 crc kubenswrapper[4688]: I1001 16:26:39.755292 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" podUID="20694bb0-a1ae-4db2-b856-772db76aafd4" containerName="manager" containerID="cri-o://2a6846880ab078151e077a6ff47237c1af3948631dfc707cb26bbab7d09601eb" gracePeriod=10 Oct 01 16:26:39 crc kubenswrapper[4688]: I1001 16:26:39.942576 4688 generic.go:334] "Generic (PLEG): container finished" podID="20694bb0-a1ae-4db2-b856-772db76aafd4" containerID="2a6846880ab078151e077a6ff47237c1af3948631dfc707cb26bbab7d09601eb" exitCode=0 Oct 01 16:26:39 crc kubenswrapper[4688]: I1001 16:26:39.942660 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" event={"ID":"20694bb0-a1ae-4db2-b856-772db76aafd4","Type":"ContainerDied","Data":"2a6846880ab078151e077a6ff47237c1af3948631dfc707cb26bbab7d09601eb"} Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.279156 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.392812 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvj68\" (UniqueName: \"kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68\") pod \"20694bb0-a1ae-4db2-b856-772db76aafd4\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.392982 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert\") pod \"20694bb0-a1ae-4db2-b856-772db76aafd4\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.393019 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert\") pod \"20694bb0-a1ae-4db2-b856-772db76aafd4\" (UID: \"20694bb0-a1ae-4db2-b856-772db76aafd4\") " Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.400733 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68" (OuterVolumeSpecName: "kube-api-access-rvj68") pod "20694bb0-a1ae-4db2-b856-772db76aafd4" (UID: "20694bb0-a1ae-4db2-b856-772db76aafd4"). InnerVolumeSpecName "kube-api-access-rvj68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.401208 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "20694bb0-a1ae-4db2-b856-772db76aafd4" (UID: "20694bb0-a1ae-4db2-b856-772db76aafd4"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.408491 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "20694bb0-a1ae-4db2-b856-772db76aafd4" (UID: "20694bb0-a1ae-4db2-b856-772db76aafd4"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.495313 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.495359 4688 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20694bb0-a1ae-4db2-b856-772db76aafd4-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.495372 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvj68\" (UniqueName: \"kubernetes.io/projected/20694bb0-a1ae-4db2-b856-772db76aafd4-kube-api-access-rvj68\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.972966 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" event={"ID":"20694bb0-a1ae-4db2-b856-772db76aafd4","Type":"ContainerDied","Data":"1990c15486ee25be0c1d2ff0e298378ca548b50eb04dc3023f568b784cca86ec"} Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.973335 4688 scope.go:117] "RemoveContainer" containerID="2a6846880ab078151e077a6ff47237c1af3948631dfc707cb26bbab7d09601eb" Oct 01 16:26:40 crc kubenswrapper[4688]: I1001 16:26:40.973163 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql" Oct 01 16:26:41 crc kubenswrapper[4688]: I1001 16:26:41.025443 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 16:26:41 crc kubenswrapper[4688]: I1001 16:26:41.038681 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6b8fb644cb-dm2ql"] Oct 01 16:26:41 crc kubenswrapper[4688]: I1001 16:26:41.393624 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20694bb0-a1ae-4db2-b856-772db76aafd4" path="/var/lib/kubelet/pods/20694bb0-a1ae-4db2-b856-772db76aafd4/volumes" Oct 01 16:26:47 crc kubenswrapper[4688]: I1001 16:26:47.382590 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:26:47 crc kubenswrapper[4688]: E1001 16:26:47.384749 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.568188 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn"] Oct 01 16:26:48 crc kubenswrapper[4688]: E1001 16:26:48.568659 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20694bb0-a1ae-4db2-b856-772db76aafd4" containerName="manager" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.568676 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="20694bb0-a1ae-4db2-b856-772db76aafd4" containerName="manager" Oct 01 16:26:48 crc kubenswrapper[4688]: E1001 16:26:48.568719 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e2aa94-6e98-416b-b112-042f4444755d" containerName="webhook-server" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.568727 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e2aa94-6e98-416b-b112-042f4444755d" containerName="webhook-server" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.569021 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e2aa94-6e98-416b-b112-042f4444755d" containerName="webhook-server" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.569044 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="20694bb0-a1ae-4db2-b856-772db76aafd4" containerName="manager" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.569856 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.597625 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn"] Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.650692 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-apiservice-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.651164 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-webhook-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.651301 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khdnz\" (UniqueName: \"kubernetes.io/projected/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-kube-api-access-khdnz\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.753444 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khdnz\" (UniqueName: \"kubernetes.io/projected/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-kube-api-access-khdnz\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.753681 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-apiservice-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.753746 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-webhook-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.764279 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-webhook-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.764510 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-apiservice-cert\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.795436 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khdnz\" (UniqueName: \"kubernetes.io/projected/550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6-kube-api-access-khdnz\") pod \"metallb-operator-controller-manager-6887bf55bf-g25xn\" (UID: \"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6\") " pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.820882 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm"] Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.845081 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.848457 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm"] Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.890140 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.957802 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-apiservice-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.957885 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-webhook-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:48 crc kubenswrapper[4688]: I1001 16:26:48.957918 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdpl\" (UniqueName: \"kubernetes.io/projected/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-kube-api-access-lsdpl\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.059261 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-apiservice-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.059676 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-webhook-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.059712 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdpl\" (UniqueName: \"kubernetes.io/projected/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-kube-api-access-lsdpl\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.066062 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-webhook-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.077915 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-apiservice-cert\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.083015 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdpl\" (UniqueName: \"kubernetes.io/projected/e8ccae4e-6cb9-40e8-b56b-4033f5d0f086-kube-api-access-lsdpl\") pod \"metallb-operator-webhook-server-9b7c45d89-n5mmm\" (UID: \"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086\") " pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.161508 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.399578 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn"] Oct 01 16:26:49 crc kubenswrapper[4688]: W1001 16:26:49.665001 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8ccae4e_6cb9_40e8_b56b_4033f5d0f086.slice/crio-3e9594d99741ea824fbdd4ec7f5ddba4e56e78149e46d26320ddac25c6821997 WatchSource:0}: Error finding container 3e9594d99741ea824fbdd4ec7f5ddba4e56e78149e46d26320ddac25c6821997: Status 404 returned error can't find the container with id 3e9594d99741ea824fbdd4ec7f5ddba4e56e78149e46d26320ddac25c6821997 Oct 01 16:26:49 crc kubenswrapper[4688]: I1001 16:26:49.679053 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm"] Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.085908 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" event={"ID":"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6","Type":"ContainerStarted","Data":"e0144f0669352e9115e1f2d7738ff3427e4b25cb458fd9a71394395209cc72db"} Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.086186 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" event={"ID":"550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6","Type":"ContainerStarted","Data":"816c9a993ebffb0f7ceafff66590c9f4b3a0ec397d3e4351877d3e19e2d3357a"} Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.086224 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.089193 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" event={"ID":"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086","Type":"ContainerStarted","Data":"143fd3a4ccb9e5aa5a6a4570e3d6036457d57f48160543ae870927b701f7309d"} Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.089320 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" event={"ID":"e8ccae4e-6cb9-40e8-b56b-4033f5d0f086","Type":"ContainerStarted","Data":"3e9594d99741ea824fbdd4ec7f5ddba4e56e78149e46d26320ddac25c6821997"} Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.089626 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.115951 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" podStartSLOduration=2.115930572 podStartE2EDuration="2.115930572s" podCreationTimestamp="2025-10-01 16:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:26:50.108653804 +0000 UTC m=+2399.459293766" watchObservedRunningTime="2025-10-01 16:26:50.115930572 +0000 UTC m=+2399.466570534" Oct 01 16:26:50 crc kubenswrapper[4688]: I1001 16:26:50.131891 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" podStartSLOduration=2.131874717 podStartE2EDuration="2.131874717s" podCreationTimestamp="2025-10-01 16:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:26:50.128002776 +0000 UTC m=+2399.478642748" watchObservedRunningTime="2025-10-01 16:26:50.131874717 +0000 UTC m=+2399.482514679" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.622726 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-96tnj"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.623825 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="controller" containerID="cri-o://3ce17c0c71b6b890863cac3393333b9816e458d480af5e92d43ff177e0f40505" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.624322 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy-frr" containerID="cri-o://ca7b92be6bbd3fefceb16c315dbd5a319f5a5109bc38beb42a83d64eb24ef1a4" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.624377 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr" containerID="cri-o://cd481bc67fef91619ada7075f2e5a399338400fa039814f49a1679bb1d4845e4" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.624385 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="reloader" containerID="cri-o://2adff539b744d26bf47f3bf8345c86fa4837bdf3d800287b224f0985f224248b" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.624538 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy" containerID="cri-o://b4e6fc1a1be0dd3cdd6e66cfd1f47397c89fa6d8f46db6a904454ee9f741fa3e" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.624603 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-96tnj" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr-metrics" containerID="cri-o://d1c486b572834d5e19902b9da70241b6656ecc730701014b1a3db045765415e5" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.656957 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/frr-k8s-96tnj"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.687121 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk"] Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.687851 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="controller" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.687969 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="controller" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.688068 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-frr-files" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.688160 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-frr-files" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.688254 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr-metrics" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.688336 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr-metrics" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.688480 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.688600 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.688704 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.688797 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.688899 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-reloader" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.689033 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-reloader" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.689134 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-metrics" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.689220 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="cp-metrics" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.689313 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="reloader" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.689404 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="reloader" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.689489 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy-frr" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.689593 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy-frr" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.689927 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.690047 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr-metrics" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.690151 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="frr" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.690246 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="reloader" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.690335 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="controller" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.690555 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" containerName="kube-rbac-proxy-frr" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.694576 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.732242 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.766169 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b047ab98-f421-44b0-9ee5-d2affd7804b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.770186 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh58x\" (UniqueName: \"kubernetes.io/projected/b047ab98-f421-44b0-9ee5-d2affd7804b8-kube-api-access-kh58x\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.793730 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pdvr5"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.798103 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872634 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872688 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md6vj\" (UniqueName: \"kubernetes.io/projected/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-kube-api-access-md6vj\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872706 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-sockets\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872791 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-startup\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872823 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh58x\" (UniqueName: \"kubernetes.io/projected/b047ab98-f421-44b0-9ee5-d2affd7804b8-kube-api-access-kh58x\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872859 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-reloader\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872898 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b047ab98-f421-44b0-9ee5-d2affd7804b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872927 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics-certs\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.872945 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-conf\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.886543 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b047ab98-f421-44b0-9ee5-d2affd7804b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.901324 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh58x\" (UniqueName: \"kubernetes.io/projected/b047ab98-f421-44b0-9ee5-d2affd7804b8-kube-api-access-kh58x\") pod \"frr-k8s-webhook-server-64bf5d555-s7mkk\" (UID: \"b047ab98-f421-44b0-9ee5-d2affd7804b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.908076 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/speaker-zpk8l"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.908888 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-zpk8l" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="speaker" containerID="cri-o://3fa460f893bab5e197bfe5d4027cdcd0e3713a87924aeb1129495be6e1247a49" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.912433 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-zpk8l" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="kube-rbac-proxy" containerID="cri-o://498d691e8ffbfc6c70c3503591c296a306aecc3c9e063777f18102b8feadf6b4" gracePeriod=2 Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.971322 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/speaker-zpk8l"] Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974485 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-reloader\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974576 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics-certs\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974595 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-conf\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974617 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974643 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md6vj\" (UniqueName: \"kubernetes.io/projected/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-kube-api-access-md6vj\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974661 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-sockets\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.974744 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-startup\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.975740 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-startup\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.975950 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-reloader\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.976673 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-conf\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.976996 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-frr-sockets\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.978646 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.990976 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-metrics-certs\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.991031 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-st9jl"] Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.991668 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.991686 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: E1001 16:26:58.991703 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="speaker" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.991709 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="speaker" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.991892 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="kube-rbac-proxy" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.991918 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerName="speaker" Oct 01 16:26:58 crc kubenswrapper[4688]: I1001 16:26:58.992826 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.027654 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-j98l6"] Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.029239 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.047720 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-st9jl"] Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.059478 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md6vj\" (UniqueName: \"kubernetes.io/projected/d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac-kube-api-access-md6vj\") pod \"frr-k8s-pdvr5\" (UID: \"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac\") " pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.079575 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-metrics-certs\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.079701 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-memberlist\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.079736 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618070a7-23e1-4b85-a68e-85196f7838d9-metallb-excludel2\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.079755 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-cert\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.079866 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-metrics-certs\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.080031 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngr4f\" (UniqueName: \"kubernetes.io/projected/57161b05-13a8-415e-b9e9-ab575ff86007-kube-api-access-ngr4f\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.080058 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwzh\" (UniqueName: \"kubernetes.io/projected/618070a7-23e1-4b85-a68e-85196f7838d9-kube-api-access-bfwzh\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.174989 4688 generic.go:334] "Generic (PLEG): container finished" podID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerID="498d691e8ffbfc6c70c3503591c296a306aecc3c9e063777f18102b8feadf6b4" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.175017 4688 generic.go:334] "Generic (PLEG): container finished" podID="1bef642c-98fe-413c-8a6d-a7a308d76036" containerID="3fa460f893bab5e197bfe5d4027cdcd0e3713a87924aeb1129495be6e1247a49" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.183431 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngr4f\" (UniqueName: \"kubernetes.io/projected/57161b05-13a8-415e-b9e9-ab575ff86007-kube-api-access-ngr4f\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.183476 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwzh\" (UniqueName: \"kubernetes.io/projected/618070a7-23e1-4b85-a68e-85196f7838d9-kube-api-access-bfwzh\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.184015 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-metrics-certs\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.184092 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-memberlist\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.184119 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618070a7-23e1-4b85-a68e-85196f7838d9-metallb-excludel2\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.184144 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-cert\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.184228 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-metrics-certs\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.185142 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618070a7-23e1-4b85-a68e-85196f7838d9-metallb-excludel2\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186202 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="ca7b92be6bbd3fefceb16c315dbd5a319f5a5109bc38beb42a83d64eb24ef1a4" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186337 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="b4e6fc1a1be0dd3cdd6e66cfd1f47397c89fa6d8f46db6a904454ee9f741fa3e" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186453 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="d1c486b572834d5e19902b9da70241b6656ecc730701014b1a3db045765415e5" exitCode=143 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186589 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="2adff539b744d26bf47f3bf8345c86fa4837bdf3d800287b224f0985f224248b" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186702 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="cd481bc67fef91619ada7075f2e5a399338400fa039814f49a1679bb1d4845e4" exitCode=143 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186799 4688 generic.go:334] "Generic (PLEG): container finished" podID="a9d78e63-3534-48cc-9df2-b986f6897909" containerID="3ce17c0c71b6b890863cac3393333b9816e458d480af5e92d43ff177e0f40505" exitCode=0 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.186933 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1638e47cf2b1656251a30986adf4d7476de48183582e697bf26fb630765a85e2" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.187472 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.189698 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-memberlist\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.190136 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618070a7-23e1-4b85-a68e-85196f7838d9-metrics-certs\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.191055 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-metrics-certs\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.197248 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57161b05-13a8-415e-b9e9-ab575ff86007-cert\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.199859 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-9b7c45d89-n5mmm" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.205207 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwzh\" (UniqueName: \"kubernetes.io/projected/618070a7-23e1-4b85-a68e-85196f7838d9-kube-api-access-bfwzh\") pod \"speaker-j98l6\" (UID: \"618070a7-23e1-4b85-a68e-85196f7838d9\") " pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.210948 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngr4f\" (UniqueName: \"kubernetes.io/projected/57161b05-13a8-415e-b9e9-ab575ff86007-kube-api-access-ngr4f\") pod \"controller-68d546b9d8-st9jl\" (UID: \"57161b05-13a8-415e-b9e9-ab575ff86007\") " pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.262434 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.335438 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-webhook-server-756f8755cd-ml958"] Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.335721 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" podUID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" containerName="webhook-server" containerID="cri-o://30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8" gracePeriod=2 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.355911 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96tnj" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.372105 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-webhook-server-756f8755cd-ml958"] Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.394824 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.411183 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-j98l6" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.489876 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.489950 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.490018 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.490127 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.490148 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.490203 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.490233 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gdk6\" (UniqueName: \"kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6\") pod \"a9d78e63-3534-48cc-9df2-b986f6897909\" (UID: \"a9d78e63-3534-48cc-9df2-b986f6897909\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.497571 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader" (OuterVolumeSpecName: "reloader") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "reloader". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.499437 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf" (OuterVolumeSpecName: "frr-conf") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "frr-conf". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.500046 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics" (OuterVolumeSpecName: "metrics") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.501144 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup" (OuterVolumeSpecName: "frr-startup") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "frr-startup". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: W1001 16:26:59.502098 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod618070a7_23e1_4b85_a68e_85196f7838d9.slice/crio-2aca539e93e2340f997cc664860dded590f8196554abc43ec73663b6805f6de3 WatchSource:0}: Error finding container 2aca539e93e2340f997cc664860dded590f8196554abc43ec73663b6805f6de3: Status 404 returned error can't find the container with id 2aca539e93e2340f997cc664860dded590f8196554abc43ec73663b6805f6de3 Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.513001 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6" (OuterVolumeSpecName: "kube-api-access-4gdk6") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "kube-api-access-4gdk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.518312 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets" (OuterVolumeSpecName: "frr-sockets") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "frr-sockets". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.520376 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "a9d78e63-3534-48cc-9df2-b986f6897909" (UID: "a9d78e63-3534-48cc-9df2-b986f6897909"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.595972 4688 reconciler_common.go:293] "Volume detached for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a9d78e63-3534-48cc-9df2-b986f6897909-frr-startup\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596328 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9d78e63-3534-48cc-9df2-b986f6897909-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596346 4688 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596359 4688 reconciler_common.go:293] "Volume detached for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-conf\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596370 4688 reconciler_common.go:293] "Volume detached for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-frr-sockets\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596380 4688 reconciler_common.go:293] "Volume detached for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a9d78e63-3534-48cc-9df2-b986f6897909-reloader\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.596391 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gdk6\" (UniqueName: \"kubernetes.io/projected/a9d78e63-3534-48cc-9df2-b986f6897909-kube-api-access-4gdk6\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.617207 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zpk8l" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.697340 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") pod \"1bef642c-98fe-413c-8a6d-a7a308d76036\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.697442 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") pod \"1bef642c-98fe-413c-8a6d-a7a308d76036\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.697497 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2\") pod \"1bef642c-98fe-413c-8a6d-a7a308d76036\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.697682 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vrk4\" (UniqueName: \"kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4\") pod \"1bef642c-98fe-413c-8a6d-a7a308d76036\" (UID: \"1bef642c-98fe-413c-8a6d-a7a308d76036\") " Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.703303 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "1bef642c-98fe-413c-8a6d-a7a308d76036" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.704039 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2" (OuterVolumeSpecName: "metallb-excludel2") pod "1bef642c-98fe-413c-8a6d-a7a308d76036" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036"). InnerVolumeSpecName "metallb-excludel2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.706954 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4" (OuterVolumeSpecName: "kube-api-access-4vrk4") pod "1bef642c-98fe-413c-8a6d-a7a308d76036" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036"). InnerVolumeSpecName "kube-api-access-4vrk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.708858 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist" (OuterVolumeSpecName: "memberlist") pod "1bef642c-98fe-413c-8a6d-a7a308d76036" (UID: "1bef642c-98fe-413c-8a6d-a7a308d76036"). InnerVolumeSpecName "memberlist". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.800117 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vrk4\" (UniqueName: \"kubernetes.io/projected/1bef642c-98fe-413c-8a6d-a7a308d76036-kube-api-access-4vrk4\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.800151 4688 reconciler_common.go:293] "Volume detached for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-memberlist\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.800162 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bef642c-98fe-413c-8a6d-a7a308d76036-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.800170 4688 reconciler_common.go:293] "Volume detached for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bef642c-98fe-413c-8a6d-a7a308d76036-metallb-excludel2\") on node \"crc\" DevicePath \"\"" Oct 01 16:26:59 crc kubenswrapper[4688]: I1001 16:26:59.948651 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk"] Oct 01 16:26:59 crc kubenswrapper[4688]: W1001 16:26:59.958657 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb047ab98_f421_44b0_9ee5_d2affd7804b8.slice/crio-60fdae9b3ebd4b0e04479cea0e0ba1aa86ce4b1241e88166f8612f13a4c13e10 WatchSource:0}: Error finding container 60fdae9b3ebd4b0e04479cea0e0ba1aa86ce4b1241e88166f8612f13a4c13e10: Status 404 returned error can't find the container with id 60fdae9b3ebd4b0e04479cea0e0ba1aa86ce4b1241e88166f8612f13a4c13e10 Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.068476 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-st9jl"] Oct 01 16:27:00 crc kubenswrapper[4688]: W1001 16:27:00.086799 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57161b05_13a8_415e_b9e9_ab575ff86007.slice/crio-579936365ddcf27fe7ca38ca972d6c7fcb9972a3ce5e71dcb0d36043dcf534e8 WatchSource:0}: Error finding container 579936365ddcf27fe7ca38ca972d6c7fcb9972a3ce5e71dcb0d36043dcf534e8: Status 404 returned error can't find the container with id 579936365ddcf27fe7ca38ca972d6c7fcb9972a3ce5e71dcb0d36043dcf534e8 Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.175300 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.202016 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"c14135117ed050fd2f30d9bee5fd3c3310886eeaac9117ef84165ae1fb69d20c"} Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.203359 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" event={"ID":"b047ab98-f421-44b0-9ee5-d2affd7804b8","Type":"ContainerStarted","Data":"60fdae9b3ebd4b0e04479cea0e0ba1aa86ce4b1241e88166f8612f13a4c13e10"} Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.205662 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-j98l6" event={"ID":"618070a7-23e1-4b85-a68e-85196f7838d9","Type":"ContainerStarted","Data":"7c0fecf1ccf15ad15b4f2759b0d7280ef25e3fec492425ce13f451f4ab6e55e6"} Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.205692 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-j98l6" event={"ID":"618070a7-23e1-4b85-a68e-85196f7838d9","Type":"ContainerStarted","Data":"2aca539e93e2340f997cc664860dded590f8196554abc43ec73663b6805f6de3"} Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.208386 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-st9jl" event={"ID":"57161b05-13a8-415e-b9e9-ab575ff86007","Type":"ContainerStarted","Data":"579936365ddcf27fe7ca38ca972d6c7fcb9972a3ce5e71dcb0d36043dcf534e8"} Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.209846 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert\") pod \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.210076 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert\") pod \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.210259 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5z2c\" (UniqueName: \"kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c\") pod \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\" (UID: \"23ec7644-389e-4fb6-977a-a1a7c69b3d7e\") " Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.214562 4688 scope.go:117] "RemoveContainer" containerID="498d691e8ffbfc6c70c3503591c296a306aecc3c9e063777f18102b8feadf6b4" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.214727 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zpk8l" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.217752 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "23ec7644-389e-4fb6-977a-a1a7c69b3d7e" (UID: "23ec7644-389e-4fb6-977a-a1a7c69b3d7e"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.221246 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "23ec7644-389e-4fb6-977a-a1a7c69b3d7e" (UID: "23ec7644-389e-4fb6-977a-a1a7c69b3d7e"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.229924 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c" (OuterVolumeSpecName: "kube-api-access-b5z2c") pod "23ec7644-389e-4fb6-977a-a1a7c69b3d7e" (UID: "23ec7644-389e-4fb6-977a-a1a7c69b3d7e"). InnerVolumeSpecName "kube-api-access-b5z2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.234788 4688 generic.go:334] "Generic (PLEG): container finished" podID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" containerID="30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8" exitCode=0 Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.234939 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96tnj" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.235869 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-756f8755cd-ml958" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.305385 4688 scope.go:117] "RemoveContainer" containerID="3fa460f893bab5e197bfe5d4027cdcd0e3713a87924aeb1129495be6e1247a49" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.316757 4688 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.316789 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.316799 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5z2c\" (UniqueName: \"kubernetes.io/projected/23ec7644-389e-4fb6-977a-a1a7c69b3d7e-kube-api-access-b5z2c\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.343682 4688 scope.go:117] "RemoveContainer" containerID="30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.388046 4688 scope.go:117] "RemoveContainer" containerID="30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8" Oct 01 16:27:00 crc kubenswrapper[4688]: E1001 16:27:00.388631 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8\": container with ID starting with 30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8 not found: ID does not exist" containerID="30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8" Oct 01 16:27:00 crc kubenswrapper[4688]: I1001 16:27:00.388661 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8"} err="failed to get container status \"30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8\": rpc error: code = NotFound desc = could not find container \"30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8\": container with ID starting with 30b0002137a013145b30822d4534bd359494c4c400e79ef0babe0338e22982b8 not found: ID does not exist" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.246703 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-j98l6" event={"ID":"618070a7-23e1-4b85-a68e-85196f7838d9","Type":"ContainerStarted","Data":"fe9e9a26f9e90db58981496567413192c0ee1150fcedd239bdd2e3ba1fd930b1"} Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.247611 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-j98l6" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.250055 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-st9jl" event={"ID":"57161b05-13a8-415e-b9e9-ab575ff86007","Type":"ContainerStarted","Data":"c1b72bb0b0ee54ec3a6fba53f47254c86ab20915c235f128c3316a129720f11e"} Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.250087 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-st9jl" event={"ID":"57161b05-13a8-415e-b9e9-ab575ff86007","Type":"ContainerStarted","Data":"da9d1a84087b504ab172a97b9891fe4ae647a6f41a083256a9ea6d3e86975389"} Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.250221 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.266776 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-j98l6" podStartSLOduration=3.266755386 podStartE2EDuration="3.266755386s" podCreationTimestamp="2025-10-01 16:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:27:01.265772028 +0000 UTC m=+2410.616412000" watchObservedRunningTime="2025-10-01 16:27:01.266755386 +0000 UTC m=+2410.617395348" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.280646 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-st9jl" podStartSLOduration=3.280623302 podStartE2EDuration="3.280623302s" podCreationTimestamp="2025-10-01 16:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:27:01.279598912 +0000 UTC m=+2410.630238904" watchObservedRunningTime="2025-10-01 16:27:01.280623302 +0000 UTC m=+2410.631263274" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.402207 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:27:01 crc kubenswrapper[4688]: E1001 16:27:01.402498 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.403090 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bef642c-98fe-413c-8a6d-a7a308d76036" path="/var/lib/kubelet/pods/1bef642c-98fe-413c-8a6d-a7a308d76036/volumes" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.405259 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" path="/var/lib/kubelet/pods/23ec7644-389e-4fb6-977a-a1a7c69b3d7e/volumes" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.406982 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d78e63-3534-48cc-9df2-b986f6897909" path="/var/lib/kubelet/pods/a9d78e63-3534-48cc-9df2-b986f6897909/volumes" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.915397 4688 scope.go:117] "RemoveContainer" containerID="cd481bc67fef91619ada7075f2e5a399338400fa039814f49a1679bb1d4845e4" Oct 01 16:27:01 crc kubenswrapper[4688]: I1001 16:27:01.981409 4688 scope.go:117] "RemoveContainer" containerID="2adff539b744d26bf47f3bf8345c86fa4837bdf3d800287b224f0985f224248b" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.009274 4688 scope.go:117] "RemoveContainer" containerID="3ce17c0c71b6b890863cac3393333b9816e458d480af5e92d43ff177e0f40505" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.053356 4688 scope.go:117] "RemoveContainer" containerID="ad383a58fb88d5d9811b0ea386b54a8aeb0c6d75709080d2b393b1139ec7eea1" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.088807 4688 scope.go:117] "RemoveContainer" containerID="ca7b92be6bbd3fefceb16c315dbd5a319f5a5109bc38beb42a83d64eb24ef1a4" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.151724 4688 scope.go:117] "RemoveContainer" containerID="b4e6fc1a1be0dd3cdd6e66cfd1f47397c89fa6d8f46db6a904454ee9f741fa3e" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.198510 4688 scope.go:117] "RemoveContainer" containerID="d1c486b572834d5e19902b9da70241b6656ecc730701014b1a3db045765415e5" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.233612 4688 scope.go:117] "RemoveContainer" containerID="8e380136f2b94fb2686295d273c4d06283e596f9754b8601e2359f68d39790ab" Oct 01 16:27:02 crc kubenswrapper[4688]: I1001 16:27:02.266505 4688 scope.go:117] "RemoveContainer" containerID="5abf2f107eda6674a107691f4534b62321e321947f60b5f7e1b84bfc9dda4719" Oct 01 16:27:08 crc kubenswrapper[4688]: I1001 16:27:08.352186 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" event={"ID":"b047ab98-f421-44b0-9ee5-d2affd7804b8","Type":"ContainerStarted","Data":"7cbdbbcc2ee7cf85e93a8d00126307e4a4ba939046a0e7907022d27233a9aaa2"} Oct 01 16:27:08 crc kubenswrapper[4688]: I1001 16:27:08.352811 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:27:08 crc kubenswrapper[4688]: I1001 16:27:08.357983 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac" containerID="7f52d380500b0da5bcebdd60127b284e8cf4215cf28a1cf89df170c78807b962" exitCode=0 Oct 01 16:27:08 crc kubenswrapper[4688]: I1001 16:27:08.358029 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerDied","Data":"7f52d380500b0da5bcebdd60127b284e8cf4215cf28a1cf89df170c78807b962"} Oct 01 16:27:08 crc kubenswrapper[4688]: I1001 16:27:08.382372 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" podStartSLOduration=3.039695621 podStartE2EDuration="10.382348961s" podCreationTimestamp="2025-10-01 16:26:58 +0000 UTC" firstStartedPulling="2025-10-01 16:26:59.973626118 +0000 UTC m=+2409.324266080" lastFinishedPulling="2025-10-01 16:27:07.316279458 +0000 UTC m=+2416.666919420" observedRunningTime="2025-10-01 16:27:08.377032909 +0000 UTC m=+2417.727672881" watchObservedRunningTime="2025-10-01 16:27:08.382348961 +0000 UTC m=+2417.732988963" Oct 01 16:27:09 crc kubenswrapper[4688]: I1001 16:27:09.368484 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac" containerID="2a7be23dc5bacffdfdb5c30cb4e48e5adfba955cfc61284ad355c337eb7820ae" exitCode=0 Oct 01 16:27:09 crc kubenswrapper[4688]: I1001 16:27:09.368609 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerDied","Data":"2a7be23dc5bacffdfdb5c30cb4e48e5adfba955cfc61284ad355c337eb7820ae"} Oct 01 16:27:09 crc kubenswrapper[4688]: I1001 16:27:09.420903 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-j98l6" Oct 01 16:27:10 crc kubenswrapper[4688]: I1001 16:27:10.379459 4688 generic.go:334] "Generic (PLEG): container finished" podID="d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac" containerID="ba57522ccce14b9cc19d3f14d746728e875ca49dcf99af41f3f4828521c7bbec" exitCode=0 Oct 01 16:27:10 crc kubenswrapper[4688]: I1001 16:27:10.379507 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerDied","Data":"ba57522ccce14b9cc19d3f14d746728e875ca49dcf99af41f3f4828521c7bbec"} Oct 01 16:27:11 crc kubenswrapper[4688]: I1001 16:27:11.404927 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"942110e502542389ff9965d1f23029b0bf9751b0bdefb06dc6c2f47a812c7fb6"} Oct 01 16:27:11 crc kubenswrapper[4688]: I1001 16:27:11.405271 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"54122f930c2a52119c95c4cf0039f16c27ee345434cbd344a73909817296d3a4"} Oct 01 16:27:11 crc kubenswrapper[4688]: I1001 16:27:11.405285 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"e57ef93a2d64754632041a5c4e15e2ef64e06022a6b0c3625f855e8c33287f7c"} Oct 01 16:27:11 crc kubenswrapper[4688]: I1001 16:27:11.405295 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"afb897c8192e506f25981974a8a76477c94d0b7dc85153996b0d40118551949c"} Oct 01 16:27:11 crc kubenswrapper[4688]: I1001 16:27:11.405305 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"9b6161abe53570a2bc3fd149ec5e447a24279dc75f5e465973ba9461b443349d"} Oct 01 16:27:12 crc kubenswrapper[4688]: I1001 16:27:12.381489 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:27:12 crc kubenswrapper[4688]: E1001 16:27:12.382078 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:27:12 crc kubenswrapper[4688]: I1001 16:27:12.416802 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pdvr5" event={"ID":"d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac","Type":"ContainerStarted","Data":"81ada84d14616147ff9a7a9a7cc32e0fecf0637af60f771e03e993bf0d604bff"} Oct 01 16:27:12 crc kubenswrapper[4688]: I1001 16:27:12.418163 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:27:12 crc kubenswrapper[4688]: I1001 16:27:12.471856 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pdvr5" podStartSLOduration=6.658015858 podStartE2EDuration="14.471833096s" podCreationTimestamp="2025-10-01 16:26:58 +0000 UTC" firstStartedPulling="2025-10-01 16:26:59.51121666 +0000 UTC m=+2408.861856622" lastFinishedPulling="2025-10-01 16:27:07.325033898 +0000 UTC m=+2416.675673860" observedRunningTime="2025-10-01 16:27:12.457920729 +0000 UTC m=+2421.808560691" watchObservedRunningTime="2025-10-01 16:27:12.471833096 +0000 UTC m=+2421.822473058" Oct 01 16:27:14 crc kubenswrapper[4688]: I1001 16:27:14.263892 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:27:14 crc kubenswrapper[4688]: I1001 16:27:14.302894 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.191349 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-s7mkk" Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.246395 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.246610 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" podUID="86cf7b4d-c977-4016-b726-57e5330a40fb" containerName="frr-k8s-webhook-server" containerID="cri-o://5d1ea6733116b3f7ae95799de9b2055125378f9095ff12acaf64e7a10497f1be" gracePeriod=10 Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.400033 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-st9jl" Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.466762 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/controller-5d688f5ffc-85kzw"] Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.467120 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-5d688f5ffc-85kzw" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="kube-rbac-proxy" containerID="cri-o://03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" gracePeriod=2 Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.467287 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-5d688f5ffc-85kzw" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="controller" containerID="cri-o://6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" gracePeriod=2 Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.480920 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/controller-5d688f5ffc-85kzw"] Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.501314 4688 generic.go:334] "Generic (PLEG): container finished" podID="86cf7b4d-c977-4016-b726-57e5330a40fb" containerID="5d1ea6733116b3f7ae95799de9b2055125378f9095ff12acaf64e7a10497f1be" exitCode=0 Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.501355 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" event={"ID":"86cf7b4d-c977-4016-b726-57e5330a40fb","Type":"ContainerDied","Data":"5d1ea6733116b3f7ae95799de9b2055125378f9095ff12acaf64e7a10497f1be"} Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.851008 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.969940 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p276m\" (UniqueName: \"kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m\") pod \"86cf7b4d-c977-4016-b726-57e5330a40fb\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.970126 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") pod \"86cf7b4d-c977-4016-b726-57e5330a40fb\" (UID: \"86cf7b4d-c977-4016-b726-57e5330a40fb\") " Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.975972 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert" (OuterVolumeSpecName: "cert") pod "86cf7b4d-c977-4016-b726-57e5330a40fb" (UID: "86cf7b4d-c977-4016-b726-57e5330a40fb"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:19 crc kubenswrapper[4688]: I1001 16:27:19.977034 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m" (OuterVolumeSpecName: "kube-api-access-p276m") pod "86cf7b4d-c977-4016-b726-57e5330a40fb" (UID: "86cf7b4d-c977-4016-b726-57e5330a40fb"). InnerVolumeSpecName "kube-api-access-p276m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.002513 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.071818 4688 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86cf7b4d-c977-4016-b726-57e5330a40fb-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.071859 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p276m\" (UniqueName: \"kubernetes.io/projected/86cf7b4d-c977-4016-b726-57e5330a40fb-kube-api-access-p276m\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.173048 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg65t\" (UniqueName: \"kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t\") pod \"4990c197-fc4e-484a-8d36-7990a3f3e68a\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.173456 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") pod \"4990c197-fc4e-484a-8d36-7990a3f3e68a\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.173610 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert\") pod \"4990c197-fc4e-484a-8d36-7990a3f3e68a\" (UID: \"4990c197-fc4e-484a-8d36-7990a3f3e68a\") " Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.176395 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t" (OuterVolumeSpecName: "kube-api-access-pg65t") pod "4990c197-fc4e-484a-8d36-7990a3f3e68a" (UID: "4990c197-fc4e-484a-8d36-7990a3f3e68a"). InnerVolumeSpecName "kube-api-access-pg65t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.179708 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "4990c197-fc4e-484a-8d36-7990a3f3e68a" (UID: "4990c197-fc4e-484a-8d36-7990a3f3e68a"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.196373 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert" (OuterVolumeSpecName: "cert") pod "4990c197-fc4e-484a-8d36-7990a3f3e68a" (UID: "4990c197-fc4e-484a-8d36-7990a3f3e68a"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.292699 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg65t\" (UniqueName: \"kubernetes.io/projected/4990c197-fc4e-484a-8d36-7990a3f3e68a-kube-api-access-pg65t\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.292999 4688 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.293072 4688 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4990c197-fc4e-484a-8d36-7990a3f3e68a-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.511451 4688 generic.go:334] "Generic (PLEG): container finished" podID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerID="03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" exitCode=0 Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.511482 4688 generic.go:334] "Generic (PLEG): container finished" podID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerID="6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" exitCode=0 Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.511510 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-85kzw" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.511554 4688 scope.go:117] "RemoveContainer" containerID="03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.514410 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" event={"ID":"86cf7b4d-c977-4016-b726-57e5330a40fb","Type":"ContainerDied","Data":"2017a2efcb03a0217985bf05be879520b7b9b7aff8e77d6432c93166e9a5c212"} Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.514590 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.535903 4688 scope.go:117] "RemoveContainer" containerID="6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.560130 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.568725 4688 scope.go:117] "RemoveContainer" containerID="03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" Oct 01 16:27:20 crc kubenswrapper[4688]: E1001 16:27:20.569770 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc\": container with ID starting with 03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc not found: ID does not exist" containerID="03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.569814 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc"} err="failed to get container status \"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc\": rpc error: code = NotFound desc = could not find container \"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc\": container with ID starting with 03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc not found: ID does not exist" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.569841 4688 scope.go:117] "RemoveContainer" containerID="6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" Oct 01 16:27:20 crc kubenswrapper[4688]: E1001 16:27:20.570142 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049\": container with ID starting with 6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049 not found: ID does not exist" containerID="6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.570178 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049"} err="failed to get container status \"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049\": rpc error: code = NotFound desc = could not find container \"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049\": container with ID starting with 6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049 not found: ID does not exist" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.570200 4688 scope.go:117] "RemoveContainer" containerID="03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.570983 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-h4qwp"] Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.571236 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc"} err="failed to get container status \"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc\": rpc error: code = NotFound desc = could not find container \"03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc\": container with ID starting with 03bed6172ef2d1a156f784a581e5cbc079404ea98cb47a7764032e2afa6260bc not found: ID does not exist" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.571264 4688 scope.go:117] "RemoveContainer" containerID="6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.571502 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049"} err="failed to get container status \"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049\": rpc error: code = NotFound desc = could not find container \"6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049\": container with ID starting with 6202a936a59515d17a5cbb5a3ea8ab2e2af5a01bb3d430f9c21bc0d871ddc049 not found: ID does not exist" Oct 01 16:27:20 crc kubenswrapper[4688]: I1001 16:27:20.571532 4688 scope.go:117] "RemoveContainer" containerID="5d1ea6733116b3f7ae95799de9b2055125378f9095ff12acaf64e7a10497f1be" Oct 01 16:27:21 crc kubenswrapper[4688]: I1001 16:27:21.396064 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" path="/var/lib/kubelet/pods/4990c197-fc4e-484a-8d36-7990a3f3e68a/volumes" Oct 01 16:27:21 crc kubenswrapper[4688]: I1001 16:27:21.397473 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86cf7b4d-c977-4016-b726-57e5330a40fb" path="/var/lib/kubelet/pods/86cf7b4d-c977-4016-b726-57e5330a40fb/volumes" Oct 01 16:27:27 crc kubenswrapper[4688]: I1001 16:27:27.380405 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:27:27 crc kubenswrapper[4688]: E1001 16:27:27.381079 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:27:28 crc kubenswrapper[4688]: I1001 16:27:28.892337 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6887bf55bf-g25xn" Oct 01 16:27:28 crc kubenswrapper[4688]: I1001 16:27:28.972122 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:27:28 crc kubenswrapper[4688]: I1001 16:27:28.972934 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" podUID="478d5733-100a-438f-84f6-ecd0a041f7ce" containerName="manager" containerID="cri-o://ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323" gracePeriod=10 Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.273930 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pdvr5" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.483437 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.571671 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert\") pod \"478d5733-100a-438f-84f6-ecd0a041f7ce\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.571812 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert\") pod \"478d5733-100a-438f-84f6-ecd0a041f7ce\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.571912 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzvmn\" (UniqueName: \"kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn\") pod \"478d5733-100a-438f-84f6-ecd0a041f7ce\" (UID: \"478d5733-100a-438f-84f6-ecd0a041f7ce\") " Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.577635 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn" (OuterVolumeSpecName: "kube-api-access-fzvmn") pod "478d5733-100a-438f-84f6-ecd0a041f7ce" (UID: "478d5733-100a-438f-84f6-ecd0a041f7ce"). InnerVolumeSpecName "kube-api-access-fzvmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.578038 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "478d5733-100a-438f-84f6-ecd0a041f7ce" (UID: "478d5733-100a-438f-84f6-ecd0a041f7ce"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.578287 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "478d5733-100a-438f-84f6-ecd0a041f7ce" (UID: "478d5733-100a-438f-84f6-ecd0a041f7ce"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.607859 4688 generic.go:334] "Generic (PLEG): container finished" podID="478d5733-100a-438f-84f6-ecd0a041f7ce" containerID="ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323" exitCode=0 Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.607900 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" event={"ID":"478d5733-100a-438f-84f6-ecd0a041f7ce","Type":"ContainerDied","Data":"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323"} Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.607926 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.607949 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl" event={"ID":"478d5733-100a-438f-84f6-ecd0a041f7ce","Type":"ContainerDied","Data":"b984232ec35d4c044864753d371b8faac94a622a831e195ded08d79b126426bb"} Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.607972 4688 scope.go:117] "RemoveContainer" containerID="ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.674022 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzvmn\" (UniqueName: \"kubernetes.io/projected/478d5733-100a-438f-84f6-ecd0a041f7ce-kube-api-access-fzvmn\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.674051 4688 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.674061 4688 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/478d5733-100a-438f-84f6-ecd0a041f7ce-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.699585 4688 scope.go:117] "RemoveContainer" containerID="ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323" Oct 01 16:27:29 crc kubenswrapper[4688]: E1001 16:27:29.700000 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323\": container with ID starting with ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323 not found: ID does not exist" containerID="ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.700060 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323"} err="failed to get container status \"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323\": rpc error: code = NotFound desc = could not find container \"ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323\": container with ID starting with ec9168a0cc5b5bbca888c90be88ace5873fe8be18df559c8b868301065452323 not found: ID does not exist" Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.700372 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:27:29 crc kubenswrapper[4688]: I1001 16:27:29.710884 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6498d9c794-fhhrl"] Oct 01 16:27:31 crc kubenswrapper[4688]: I1001 16:27:31.393693 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="478d5733-100a-438f-84f6-ecd0a041f7ce" path="/var/lib/kubelet/pods/478d5733-100a-438f-84f6-ecd0a041f7ce/volumes" Oct 01 16:27:41 crc kubenswrapper[4688]: I1001 16:27:41.388001 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:27:41 crc kubenswrapper[4688]: E1001 16:27:41.390600 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:27:53 crc kubenswrapper[4688]: I1001 16:27:53.833564 4688 generic.go:334] "Generic (PLEG): container finished" podID="959f38f8-3df7-422b-9c09-a6fa68d4b8f0" containerID="5aac6b7727d74273534b0c1232bac6e4e21f8d5d4ad73c93fe1a32200724fc11" exitCode=0 Oct 01 16:27:53 crc kubenswrapper[4688]: I1001 16:27:53.833633 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" event={"ID":"959f38f8-3df7-422b-9c09-a6fa68d4b8f0","Type":"ContainerDied","Data":"5aac6b7727d74273534b0c1232bac6e4e21f8d5d4ad73c93fe1a32200724fc11"} Oct 01 16:27:54 crc kubenswrapper[4688]: I1001 16:27:54.380443 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:27:54 crc kubenswrapper[4688]: E1001 16:27:54.381084 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.311343 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.459746 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0\") pod \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.459867 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd4k8\" (UniqueName: \"kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8\") pod \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.460063 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key\") pod \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.460142 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory\") pod \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.460182 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle\") pod \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\" (UID: \"959f38f8-3df7-422b-9c09-a6fa68d4b8f0\") " Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.465817 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "959f38f8-3df7-422b-9c09-a6fa68d4b8f0" (UID: "959f38f8-3df7-422b-9c09-a6fa68d4b8f0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.466729 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8" (OuterVolumeSpecName: "kube-api-access-sd4k8") pod "959f38f8-3df7-422b-9c09-a6fa68d4b8f0" (UID: "959f38f8-3df7-422b-9c09-a6fa68d4b8f0"). InnerVolumeSpecName "kube-api-access-sd4k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.491569 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "959f38f8-3df7-422b-9c09-a6fa68d4b8f0" (UID: "959f38f8-3df7-422b-9c09-a6fa68d4b8f0"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.494906 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory" (OuterVolumeSpecName: "inventory") pod "959f38f8-3df7-422b-9c09-a6fa68d4b8f0" (UID: "959f38f8-3df7-422b-9c09-a6fa68d4b8f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.497666 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "959f38f8-3df7-422b-9c09-a6fa68d4b8f0" (UID: "959f38f8-3df7-422b-9c09-a6fa68d4b8f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.563544 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.563586 4688 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.563602 4688 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.563610 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd4k8\" (UniqueName: \"kubernetes.io/projected/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-kube-api-access-sd4k8\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.563620 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/959f38f8-3df7-422b-9c09-a6fa68d4b8f0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.860119 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" event={"ID":"959f38f8-3df7-422b-9c09-a6fa68d4b8f0","Type":"ContainerDied","Data":"39bc71e1f2d3638bd5671e0504c956fcf6e11371333939351083486afa1fa4ac"} Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.860184 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39bc71e1f2d3638bd5671e0504c956fcf6e11371333939351083486afa1fa4ac" Oct 01 16:27:55 crc kubenswrapper[4688]: I1001 16:27:55.860253 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.075193 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m"] Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076082 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86cf7b4d-c977-4016-b726-57e5330a40fb" containerName="frr-k8s-webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076110 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="86cf7b4d-c977-4016-b726-57e5330a40fb" containerName="frr-k8s-webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076148 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="controller" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076161 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="controller" Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076182 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959f38f8-3df7-422b-9c09-a6fa68d4b8f0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076208 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="959f38f8-3df7-422b-9c09-a6fa68d4b8f0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076240 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="kube-rbac-proxy" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076253 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="kube-rbac-proxy" Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076270 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478d5733-100a-438f-84f6-ecd0a041f7ce" containerName="manager" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076281 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="478d5733-100a-438f-84f6-ecd0a041f7ce" containerName="manager" Oct 01 16:27:56 crc kubenswrapper[4688]: E1001 16:27:56.076325 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" containerName="webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076337 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" containerName="webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076674 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="959f38f8-3df7-422b-9c09-a6fa68d4b8f0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076707 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="86cf7b4d-c977-4016-b726-57e5330a40fb" containerName="frr-k8s-webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076726 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="controller" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076744 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="4990c197-fc4e-484a-8d36-7990a3f3e68a" containerName="kube-rbac-proxy" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076762 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="478d5733-100a-438f-84f6-ecd0a041f7ce" containerName="manager" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.076788 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ec7644-389e-4fb6-977a-a1a7c69b3d7e" containerName="webhook-server" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.077748 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083116 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083289 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083397 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083509 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083621 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083730 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.083832 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.097121 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m"] Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.175894 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.175953 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.175974 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176021 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176050 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49dnq\" (UniqueName: \"kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176075 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176172 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176215 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.176241 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278160 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278199 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278250 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278283 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49dnq\" (UniqueName: \"kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278311 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278361 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278396 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278422 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.278455 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.280239 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.282653 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.283852 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.284007 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.285125 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.285406 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.288263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.288874 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.301965 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49dnq\" (UniqueName: \"kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8g64m\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:56 crc kubenswrapper[4688]: I1001 16:27:56.401392 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:27:57 crc kubenswrapper[4688]: I1001 16:27:56.925365 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m"] Oct 01 16:27:57 crc kubenswrapper[4688]: I1001 16:27:57.879406 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" event={"ID":"6206a3c7-1bb2-422a-8bc9-38e62d890513","Type":"ContainerStarted","Data":"03ea2062adb4a5efaad97b514705813935e7ad13aba8aa93306655ff65dd5529"} Oct 01 16:27:57 crc kubenswrapper[4688]: I1001 16:27:57.880014 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" event={"ID":"6206a3c7-1bb2-422a-8bc9-38e62d890513","Type":"ContainerStarted","Data":"c1e2e5ceb7c4aeb0ad4548f34a26f5bdeb8ae14ffc8032c6ccaeda88c89af5ad"} Oct 01 16:27:57 crc kubenswrapper[4688]: I1001 16:27:57.905099 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" podStartSLOduration=1.440429488 podStartE2EDuration="1.905076469s" podCreationTimestamp="2025-10-01 16:27:56 +0000 UTC" firstStartedPulling="2025-10-01 16:27:56.937735642 +0000 UTC m=+2466.288375604" lastFinishedPulling="2025-10-01 16:27:57.402382623 +0000 UTC m=+2466.753022585" observedRunningTime="2025-10-01 16:27:57.902767993 +0000 UTC m=+2467.253408005" watchObservedRunningTime="2025-10-01 16:27:57.905076469 +0000 UTC m=+2467.255716471" Oct 01 16:28:08 crc kubenswrapper[4688]: I1001 16:28:08.381180 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:28:08 crc kubenswrapper[4688]: E1001 16:28:08.382929 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:28:23 crc kubenswrapper[4688]: I1001 16:28:23.385215 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:28:23 crc kubenswrapper[4688]: E1001 16:28:23.386462 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:28:35 crc kubenswrapper[4688]: I1001 16:28:35.381706 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:28:36 crc kubenswrapper[4688]: I1001 16:28:36.272085 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1"} Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.157999 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59"] Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.160980 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.163989 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.164253 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.185209 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59"] Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.229465 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.229516 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.229625 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rnxx\" (UniqueName: \"kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.331957 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.332017 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.332058 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rnxx\" (UniqueName: \"kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.332942 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.348367 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.350661 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rnxx\" (UniqueName: \"kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx\") pod \"collect-profiles-29322270-7zt59\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.481775 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:00 crc kubenswrapper[4688]: I1001 16:30:00.940946 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59"] Oct 01 16:30:01 crc kubenswrapper[4688]: I1001 16:30:01.110765 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" event={"ID":"9cb9259b-4194-472c-883e-13c3610ace7f","Type":"ContainerStarted","Data":"533404decceca01ef20b72fb2f68eccac31fb2be34c572d52206a92921e8387b"} Oct 01 16:30:02 crc kubenswrapper[4688]: I1001 16:30:02.123562 4688 generic.go:334] "Generic (PLEG): container finished" podID="9cb9259b-4194-472c-883e-13c3610ace7f" containerID="0aad218d2dc69829303f465f6a7ba3d1879f28485b0c82748311849f0b485e2d" exitCode=0 Oct 01 16:30:02 crc kubenswrapper[4688]: I1001 16:30:02.123612 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" event={"ID":"9cb9259b-4194-472c-883e-13c3610ace7f","Type":"ContainerDied","Data":"0aad218d2dc69829303f465f6a7ba3d1879f28485b0c82748311849f0b485e2d"} Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.496367 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.602595 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rnxx\" (UniqueName: \"kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx\") pod \"9cb9259b-4194-472c-883e-13c3610ace7f\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.602958 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume\") pod \"9cb9259b-4194-472c-883e-13c3610ace7f\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.603251 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume\") pod \"9cb9259b-4194-472c-883e-13c3610ace7f\" (UID: \"9cb9259b-4194-472c-883e-13c3610ace7f\") " Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.604141 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume" (OuterVolumeSpecName: "config-volume") pod "9cb9259b-4194-472c-883e-13c3610ace7f" (UID: "9cb9259b-4194-472c-883e-13c3610ace7f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.607966 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx" (OuterVolumeSpecName: "kube-api-access-5rnxx") pod "9cb9259b-4194-472c-883e-13c3610ace7f" (UID: "9cb9259b-4194-472c-883e-13c3610ace7f"). InnerVolumeSpecName "kube-api-access-5rnxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.617886 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9cb9259b-4194-472c-883e-13c3610ace7f" (UID: "9cb9259b-4194-472c-883e-13c3610ace7f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.706047 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9cb9259b-4194-472c-883e-13c3610ace7f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.706094 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rnxx\" (UniqueName: \"kubernetes.io/projected/9cb9259b-4194-472c-883e-13c3610ace7f-kube-api-access-5rnxx\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:03 crc kubenswrapper[4688]: I1001 16:30:03.706109 4688 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9cb9259b-4194-472c-883e-13c3610ace7f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:30:04 crc kubenswrapper[4688]: I1001 16:30:04.142327 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" event={"ID":"9cb9259b-4194-472c-883e-13c3610ace7f","Type":"ContainerDied","Data":"533404decceca01ef20b72fb2f68eccac31fb2be34c572d52206a92921e8387b"} Oct 01 16:30:04 crc kubenswrapper[4688]: I1001 16:30:04.142367 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="533404decceca01ef20b72fb2f68eccac31fb2be34c572d52206a92921e8387b" Oct 01 16:30:04 crc kubenswrapper[4688]: I1001 16:30:04.142412 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322270-7zt59" Oct 01 16:30:04 crc kubenswrapper[4688]: I1001 16:30:04.580565 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9"] Oct 01 16:30:04 crc kubenswrapper[4688]: I1001 16:30:04.590050 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322225-nv5w9"] Oct 01 16:30:05 crc kubenswrapper[4688]: I1001 16:30:05.399904 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7ec1934-3a2e-4290-aa7e-581bbb58163e" path="/var/lib/kubelet/pods/b7ec1934-3a2e-4290-aa7e-581bbb58163e/volumes" Oct 01 16:30:55 crc kubenswrapper[4688]: I1001 16:30:55.053547 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:30:55 crc kubenswrapper[4688]: I1001 16:30:55.054101 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:31:02 crc kubenswrapper[4688]: I1001 16:31:02.523882 4688 scope.go:117] "RemoveContainer" containerID="adbe432effa45d81db128c2109f68cc1d3218444aacbafdbcaceef21bdc47669" Oct 01 16:31:25 crc kubenswrapper[4688]: I1001 16:31:25.053344 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:31:25 crc kubenswrapper[4688]: I1001 16:31:25.054230 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:31:29 crc kubenswrapper[4688]: I1001 16:31:29.983909 4688 generic.go:334] "Generic (PLEG): container finished" podID="6206a3c7-1bb2-422a-8bc9-38e62d890513" containerID="03ea2062adb4a5efaad97b514705813935e7ad13aba8aa93306655ff65dd5529" exitCode=0 Oct 01 16:31:29 crc kubenswrapper[4688]: I1001 16:31:29.984391 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" event={"ID":"6206a3c7-1bb2-422a-8bc9-38e62d890513","Type":"ContainerDied","Data":"03ea2062adb4a5efaad97b514705813935e7ad13aba8aa93306655ff65dd5529"} Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.411019 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.570376 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.570514 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.571573 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.571615 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.571648 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49dnq\" (UniqueName: \"kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.571704 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.572272 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.572336 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.572416 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key\") pod \"6206a3c7-1bb2-422a-8bc9-38e62d890513\" (UID: \"6206a3c7-1bb2-422a-8bc9-38e62d890513\") " Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.576517 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq" (OuterVolumeSpecName: "kube-api-access-49dnq") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "kube-api-access-49dnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.589047 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.607418 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.612009 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.636546 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.641237 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.641838 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory" (OuterVolumeSpecName: "inventory") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.649217 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.660774 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6206a3c7-1bb2-422a-8bc9-38e62d890513" (UID: "6206a3c7-1bb2-422a-8bc9-38e62d890513"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675162 4688 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675338 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675398 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49dnq\" (UniqueName: \"kubernetes.io/projected/6206a3c7-1bb2-422a-8bc9-38e62d890513-kube-api-access-49dnq\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675450 4688 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675501 4688 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675566 4688 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675617 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675675 4688 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:31 crc kubenswrapper[4688]: I1001 16:31:31.675732 4688 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6206a3c7-1bb2-422a-8bc9-38e62d890513-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.009173 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" event={"ID":"6206a3c7-1bb2-422a-8bc9-38e62d890513","Type":"ContainerDied","Data":"c1e2e5ceb7c4aeb0ad4548f34a26f5bdeb8ae14ffc8032c6ccaeda88c89af5ad"} Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.009211 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1e2e5ceb7c4aeb0ad4548f34a26f5bdeb8ae14ffc8032c6ccaeda88c89af5ad" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.009293 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8g64m" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.184570 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2"] Oct 01 16:31:32 crc kubenswrapper[4688]: E1001 16:31:32.185100 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6206a3c7-1bb2-422a-8bc9-38e62d890513" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.185124 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="6206a3c7-1bb2-422a-8bc9-38e62d890513" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:32 crc kubenswrapper[4688]: E1001 16:31:32.185158 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb9259b-4194-472c-883e-13c3610ace7f" containerName="collect-profiles" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.185166 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb9259b-4194-472c-883e-13c3610ace7f" containerName="collect-profiles" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.185420 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb9259b-4194-472c-883e-13c3610ace7f" containerName="collect-profiles" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.185447 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="6206a3c7-1bb2-422a-8bc9-38e62d890513" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.186355 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.188699 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.188839 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.189079 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sc2gb" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.189306 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.189417 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.198481 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2"] Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285548 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285627 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285685 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285713 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285742 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285831 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6tgw\" (UniqueName: \"kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.285863 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387685 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6tgw\" (UniqueName: \"kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387750 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387826 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387860 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387898 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387918 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.387938 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.392292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.392316 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.392712 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.393233 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.393927 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.395197 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.408061 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6tgw\" (UniqueName: \"kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:32 crc kubenswrapper[4688]: I1001 16:31:32.511204 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:31:33 crc kubenswrapper[4688]: I1001 16:31:33.081795 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2"] Oct 01 16:31:33 crc kubenswrapper[4688]: I1001 16:31:33.094003 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:31:34 crc kubenswrapper[4688]: I1001 16:31:34.031697 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" event={"ID":"149aba72-0ea1-4a18-9ff1-98cfd0ada453","Type":"ContainerStarted","Data":"3fa260a611506e365600a60f55fa88d90b55c4011f13415c95ebbbb80aa2fd16"} Oct 01 16:31:35 crc kubenswrapper[4688]: I1001 16:31:35.046182 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" event={"ID":"149aba72-0ea1-4a18-9ff1-98cfd0ada453","Type":"ContainerStarted","Data":"c257a2e3b848b26a554959cfdf96e5bb5083f17d13bd6bdf948c9192e26f81b2"} Oct 01 16:31:35 crc kubenswrapper[4688]: I1001 16:31:35.080176 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" podStartSLOduration=2.361190407 podStartE2EDuration="3.080143893s" podCreationTimestamp="2025-10-01 16:31:32 +0000 UTC" firstStartedPulling="2025-10-01 16:31:33.093698845 +0000 UTC m=+2682.444338827" lastFinishedPulling="2025-10-01 16:31:33.812652311 +0000 UTC m=+2683.163292313" observedRunningTime="2025-10-01 16:31:35.071811904 +0000 UTC m=+2684.422451906" watchObservedRunningTime="2025-10-01 16:31:35.080143893 +0000 UTC m=+2684.430783895" Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.052453 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.053121 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.053167 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.053969 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.054042 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1" gracePeriod=600 Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.257904 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1" exitCode=0 Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.257947 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1"} Oct 01 16:31:55 crc kubenswrapper[4688]: I1001 16:31:55.258002 4688 scope.go:117] "RemoveContainer" containerID="4ce7013ef580052610b5ae43b04a4831d25456c3a1b47e946fe5ba1750b1bf5b" Oct 01 16:31:56 crc kubenswrapper[4688]: I1001 16:31:56.271463 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b"} Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.726433 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.729715 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.746827 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.809923 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv88g\" (UniqueName: \"kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.810334 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.810452 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.912900 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.913122 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv88g\" (UniqueName: \"kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.913172 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.913502 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.913826 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:34 crc kubenswrapper[4688]: I1001 16:32:34.936859 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv88g\" (UniqueName: \"kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g\") pod \"certified-operators-mz8m2\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:35 crc kubenswrapper[4688]: I1001 16:32:35.061321 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:35 crc kubenswrapper[4688]: I1001 16:32:35.661427 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:35 crc kubenswrapper[4688]: I1001 16:32:35.692707 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerStarted","Data":"6f929377a0ed5d66ad3a13be905aaa86e040cd597cab1722de3c475a7430b99d"} Oct 01 16:32:36 crc kubenswrapper[4688]: I1001 16:32:36.709384 4688 generic.go:334] "Generic (PLEG): container finished" podID="773b3137-2c72-460f-9505-eb2a45c107fc" containerID="de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d" exitCode=0 Oct 01 16:32:36 crc kubenswrapper[4688]: I1001 16:32:36.709573 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerDied","Data":"de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d"} Oct 01 16:32:37 crc kubenswrapper[4688]: I1001 16:32:37.724965 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerStarted","Data":"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409"} Oct 01 16:32:39 crc kubenswrapper[4688]: I1001 16:32:39.753267 4688 generic.go:334] "Generic (PLEG): container finished" podID="773b3137-2c72-460f-9505-eb2a45c107fc" containerID="3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409" exitCode=0 Oct 01 16:32:39 crc kubenswrapper[4688]: I1001 16:32:39.753388 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerDied","Data":"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409"} Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.276700 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.279545 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.331536 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.355573 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.355630 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.355685 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swssm\" (UniqueName: \"kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.457403 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.457721 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.457889 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swssm\" (UniqueName: \"kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.457991 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.458176 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.486566 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swssm\" (UniqueName: \"kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm\") pod \"redhat-marketplace-wb92w\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.614188 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:40 crc kubenswrapper[4688]: I1001 16:32:40.765440 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerStarted","Data":"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8"} Oct 01 16:32:41 crc kubenswrapper[4688]: I1001 16:32:41.100748 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mz8m2" podStartSLOduration=3.571926625 podStartE2EDuration="7.100716028s" podCreationTimestamp="2025-10-01 16:32:34 +0000 UTC" firstStartedPulling="2025-10-01 16:32:36.712286044 +0000 UTC m=+2746.062926006" lastFinishedPulling="2025-10-01 16:32:40.241075447 +0000 UTC m=+2749.591715409" observedRunningTime="2025-10-01 16:32:40.797894643 +0000 UTC m=+2750.148534605" watchObservedRunningTime="2025-10-01 16:32:41.100716028 +0000 UTC m=+2750.451355990" Oct 01 16:32:41 crc kubenswrapper[4688]: I1001 16:32:41.103600 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:41 crc kubenswrapper[4688]: I1001 16:32:41.777035 4688 generic.go:334] "Generic (PLEG): container finished" podID="a303869d-96d7-41f0-ab91-37b2505fd625" containerID="2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479" exitCode=0 Oct 01 16:32:41 crc kubenswrapper[4688]: I1001 16:32:41.777189 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerDied","Data":"2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479"} Oct 01 16:32:41 crc kubenswrapper[4688]: I1001 16:32:41.777459 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerStarted","Data":"b55f1a74fc7c1b6f187ddd54fa3ac042794ca51bc28497bca6b6540eebd15ef6"} Oct 01 16:32:43 crc kubenswrapper[4688]: I1001 16:32:43.805154 4688 generic.go:334] "Generic (PLEG): container finished" podID="a303869d-96d7-41f0-ab91-37b2505fd625" containerID="214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b" exitCode=0 Oct 01 16:32:43 crc kubenswrapper[4688]: I1001 16:32:43.805255 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerDied","Data":"214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b"} Oct 01 16:32:44 crc kubenswrapper[4688]: I1001 16:32:44.819326 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerStarted","Data":"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa"} Oct 01 16:32:45 crc kubenswrapper[4688]: I1001 16:32:45.061957 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:45 crc kubenswrapper[4688]: I1001 16:32:45.062388 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:45 crc kubenswrapper[4688]: I1001 16:32:45.117269 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:45 crc kubenswrapper[4688]: I1001 16:32:45.144019 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wb92w" podStartSLOduration=2.714381312 podStartE2EDuration="5.1439967s" podCreationTimestamp="2025-10-01 16:32:40 +0000 UTC" firstStartedPulling="2025-10-01 16:32:41.794672055 +0000 UTC m=+2751.145312017" lastFinishedPulling="2025-10-01 16:32:44.224287433 +0000 UTC m=+2753.574927405" observedRunningTime="2025-10-01 16:32:44.853607163 +0000 UTC m=+2754.204247125" watchObservedRunningTime="2025-10-01 16:32:45.1439967 +0000 UTC m=+2754.494636662" Oct 01 16:32:45 crc kubenswrapper[4688]: I1001 16:32:45.899028 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:47 crc kubenswrapper[4688]: I1001 16:32:47.269440 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:47 crc kubenswrapper[4688]: I1001 16:32:47.848986 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mz8m2" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="registry-server" containerID="cri-o://2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8" gracePeriod=2 Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.403992 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.546883 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities\") pod \"773b3137-2c72-460f-9505-eb2a45c107fc\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.547011 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv88g\" (UniqueName: \"kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g\") pod \"773b3137-2c72-460f-9505-eb2a45c107fc\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.547160 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content\") pod \"773b3137-2c72-460f-9505-eb2a45c107fc\" (UID: \"773b3137-2c72-460f-9505-eb2a45c107fc\") " Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.548189 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities" (OuterVolumeSpecName: "utilities") pod "773b3137-2c72-460f-9505-eb2a45c107fc" (UID: "773b3137-2c72-460f-9505-eb2a45c107fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.559012 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g" (OuterVolumeSpecName: "kube-api-access-wv88g") pod "773b3137-2c72-460f-9505-eb2a45c107fc" (UID: "773b3137-2c72-460f-9505-eb2a45c107fc"). InnerVolumeSpecName "kube-api-access-wv88g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.597996 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "773b3137-2c72-460f-9505-eb2a45c107fc" (UID: "773b3137-2c72-460f-9505-eb2a45c107fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.650750 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv88g\" (UniqueName: \"kubernetes.io/projected/773b3137-2c72-460f-9505-eb2a45c107fc-kube-api-access-wv88g\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.650796 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.650811 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773b3137-2c72-460f-9505-eb2a45c107fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.860830 4688 generic.go:334] "Generic (PLEG): container finished" podID="773b3137-2c72-460f-9505-eb2a45c107fc" containerID="2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8" exitCode=0 Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.860899 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerDied","Data":"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8"} Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.861189 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mz8m2" event={"ID":"773b3137-2c72-460f-9505-eb2a45c107fc","Type":"ContainerDied","Data":"6f929377a0ed5d66ad3a13be905aaa86e040cd597cab1722de3c475a7430b99d"} Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.861216 4688 scope.go:117] "RemoveContainer" containerID="2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.860990 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mz8m2" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.896116 4688 scope.go:117] "RemoveContainer" containerID="3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.896677 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.905195 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mz8m2"] Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.926011 4688 scope.go:117] "RemoveContainer" containerID="de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.977638 4688 scope.go:117] "RemoveContainer" containerID="2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8" Oct 01 16:32:48 crc kubenswrapper[4688]: E1001 16:32:48.978167 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8\": container with ID starting with 2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8 not found: ID does not exist" containerID="2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.978198 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8"} err="failed to get container status \"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8\": rpc error: code = NotFound desc = could not find container \"2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8\": container with ID starting with 2bad73b24afe244648e46139898c73fd0ed8178ac5920dd914bb929b091406c8 not found: ID does not exist" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.978222 4688 scope.go:117] "RemoveContainer" containerID="3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409" Oct 01 16:32:48 crc kubenswrapper[4688]: E1001 16:32:48.978795 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409\": container with ID starting with 3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409 not found: ID does not exist" containerID="3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.978844 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409"} err="failed to get container status \"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409\": rpc error: code = NotFound desc = could not find container \"3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409\": container with ID starting with 3f6a69476cdcd4009f05ba06d6a525e3a753e8668b1fca225cb6d9999694e409 not found: ID does not exist" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.978860 4688 scope.go:117] "RemoveContainer" containerID="de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d" Oct 01 16:32:48 crc kubenswrapper[4688]: E1001 16:32:48.979172 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d\": container with ID starting with de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d not found: ID does not exist" containerID="de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d" Oct 01 16:32:48 crc kubenswrapper[4688]: I1001 16:32:48.979197 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d"} err="failed to get container status \"de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d\": rpc error: code = NotFound desc = could not find container \"de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d\": container with ID starting with de12df2b7f19715ee87a93fe6fef0fdbe986198376ad694c915aa08e2091aa8d not found: ID does not exist" Oct 01 16:32:49 crc kubenswrapper[4688]: I1001 16:32:49.397947 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" path="/var/lib/kubelet/pods/773b3137-2c72-460f-9505-eb2a45c107fc/volumes" Oct 01 16:32:50 crc kubenswrapper[4688]: I1001 16:32:50.615888 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:50 crc kubenswrapper[4688]: I1001 16:32:50.616177 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:50 crc kubenswrapper[4688]: I1001 16:32:50.683437 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:50 crc kubenswrapper[4688]: I1001 16:32:50.944263 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:51 crc kubenswrapper[4688]: I1001 16:32:51.674992 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:52 crc kubenswrapper[4688]: I1001 16:32:52.923173 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wb92w" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="registry-server" containerID="cri-o://ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa" gracePeriod=2 Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.353008 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.373578 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content\") pod \"a303869d-96d7-41f0-ab91-37b2505fd625\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.373684 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities\") pod \"a303869d-96d7-41f0-ab91-37b2505fd625\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.373734 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swssm\" (UniqueName: \"kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm\") pod \"a303869d-96d7-41f0-ab91-37b2505fd625\" (UID: \"a303869d-96d7-41f0-ab91-37b2505fd625\") " Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.374889 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities" (OuterVolumeSpecName: "utilities") pod "a303869d-96d7-41f0-ab91-37b2505fd625" (UID: "a303869d-96d7-41f0-ab91-37b2505fd625"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.380976 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.391883 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm" (OuterVolumeSpecName: "kube-api-access-swssm") pod "a303869d-96d7-41f0-ab91-37b2505fd625" (UID: "a303869d-96d7-41f0-ab91-37b2505fd625"). InnerVolumeSpecName "kube-api-access-swssm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.395322 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a303869d-96d7-41f0-ab91-37b2505fd625" (UID: "a303869d-96d7-41f0-ab91-37b2505fd625"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.483415 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a303869d-96d7-41f0-ab91-37b2505fd625-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.483674 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swssm\" (UniqueName: \"kubernetes.io/projected/a303869d-96d7-41f0-ab91-37b2505fd625-kube-api-access-swssm\") on node \"crc\" DevicePath \"\"" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.934271 4688 generic.go:334] "Generic (PLEG): container finished" podID="a303869d-96d7-41f0-ab91-37b2505fd625" containerID="ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa" exitCode=0 Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.934322 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerDied","Data":"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa"} Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.934363 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb92w" event={"ID":"a303869d-96d7-41f0-ab91-37b2505fd625","Type":"ContainerDied","Data":"b55f1a74fc7c1b6f187ddd54fa3ac042794ca51bc28497bca6b6540eebd15ef6"} Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.934390 4688 scope.go:117] "RemoveContainer" containerID="ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.934394 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb92w" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.962743 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.966468 4688 scope.go:117] "RemoveContainer" containerID="214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b" Oct 01 16:32:53 crc kubenswrapper[4688]: I1001 16:32:53.980513 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb92w"] Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.008459 4688 scope.go:117] "RemoveContainer" containerID="2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.048017 4688 scope.go:117] "RemoveContainer" containerID="ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa" Oct 01 16:32:54 crc kubenswrapper[4688]: E1001 16:32:54.048899 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa\": container with ID starting with ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa not found: ID does not exist" containerID="ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.048955 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa"} err="failed to get container status \"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa\": rpc error: code = NotFound desc = could not find container \"ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa\": container with ID starting with ee549df59814e2c0e868a42c8c4bb8f1614975fb77368556d59d69b206da4baa not found: ID does not exist" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.049011 4688 scope.go:117] "RemoveContainer" containerID="214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b" Oct 01 16:32:54 crc kubenswrapper[4688]: E1001 16:32:54.049363 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b\": container with ID starting with 214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b not found: ID does not exist" containerID="214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.049398 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b"} err="failed to get container status \"214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b\": rpc error: code = NotFound desc = could not find container \"214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b\": container with ID starting with 214d59c29f401436633cbdd63d0e2d3144861cf3dc73d9787f19598e6b81444b not found: ID does not exist" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.049421 4688 scope.go:117] "RemoveContainer" containerID="2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479" Oct 01 16:32:54 crc kubenswrapper[4688]: E1001 16:32:54.049703 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479\": container with ID starting with 2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479 not found: ID does not exist" containerID="2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479" Oct 01 16:32:54 crc kubenswrapper[4688]: I1001 16:32:54.049784 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479"} err="failed to get container status \"2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479\": rpc error: code = NotFound desc = could not find container \"2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479\": container with ID starting with 2440d04cc92cf4e1e7b056366c71650a358c59249755e40cd3d2af8710838479 not found: ID does not exist" Oct 01 16:32:55 crc kubenswrapper[4688]: I1001 16:32:55.395094 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" path="/var/lib/kubelet/pods/a303869d-96d7-41f0-ab91-37b2505fd625/volumes" Oct 01 16:33:55 crc kubenswrapper[4688]: I1001 16:33:55.052961 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:33:55 crc kubenswrapper[4688]: I1001 16:33:55.053612 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:34:25 crc kubenswrapper[4688]: I1001 16:34:25.052517 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:34:25 crc kubenswrapper[4688]: I1001 16:34:25.054219 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:34:46 crc kubenswrapper[4688]: I1001 16:34:46.121655 4688 generic.go:334] "Generic (PLEG): container finished" podID="149aba72-0ea1-4a18-9ff1-98cfd0ada453" containerID="c257a2e3b848b26a554959cfdf96e5bb5083f17d13bd6bdf948c9192e26f81b2" exitCode=0 Oct 01 16:34:46 crc kubenswrapper[4688]: I1001 16:34:46.121753 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" event={"ID":"149aba72-0ea1-4a18-9ff1-98cfd0ada453","Type":"ContainerDied","Data":"c257a2e3b848b26a554959cfdf96e5bb5083f17d13bd6bdf948c9192e26f81b2"} Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.610595 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.776992 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6tgw\" (UniqueName: \"kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777418 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777478 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777623 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777796 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777864 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.777902 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle\") pod \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\" (UID: \"149aba72-0ea1-4a18-9ff1-98cfd0ada453\") " Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.785955 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.789134 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw" (OuterVolumeSpecName: "kube-api-access-c6tgw") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "kube-api-access-c6tgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.809513 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.812227 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory" (OuterVolumeSpecName: "inventory") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.824407 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.825803 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.842277 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "149aba72-0ea1-4a18-9ff1-98cfd0ada453" (UID: "149aba72-0ea1-4a18-9ff1-98cfd0ada453"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881021 4688 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881057 4688 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881070 4688 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881080 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6tgw\" (UniqueName: \"kubernetes.io/projected/149aba72-0ea1-4a18-9ff1-98cfd0ada453-kube-api-access-c6tgw\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881090 4688 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881101 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:47 crc kubenswrapper[4688]: I1001 16:34:47.881111 4688 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/149aba72-0ea1-4a18-9ff1-98cfd0ada453-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 01 16:34:48 crc kubenswrapper[4688]: I1001 16:34:48.151587 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" event={"ID":"149aba72-0ea1-4a18-9ff1-98cfd0ada453","Type":"ContainerDied","Data":"3fa260a611506e365600a60f55fa88d90b55c4011f13415c95ebbbb80aa2fd16"} Oct 01 16:34:48 crc kubenswrapper[4688]: I1001 16:34:48.151630 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fa260a611506e365600a60f55fa88d90b55c4011f13415c95ebbbb80aa2fd16" Oct 01 16:34:48 crc kubenswrapper[4688]: I1001 16:34:48.151706 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.053506 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.054092 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.054153 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.055102 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.055161 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" gracePeriod=600 Oct 01 16:34:55 crc kubenswrapper[4688]: E1001 16:34:55.190718 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.226282 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" exitCode=0 Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.226330 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b"} Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.226410 4688 scope.go:117] "RemoveContainer" containerID="16f5e931010b47745fc9910d96d02137df1cc10449a62f0c9e07e9345093d4c1" Oct 01 16:34:55 crc kubenswrapper[4688]: I1001 16:34:55.227209 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:34:55 crc kubenswrapper[4688]: E1001 16:34:55.227788 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:35:06 crc kubenswrapper[4688]: I1001 16:35:06.382421 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:35:06 crc kubenswrapper[4688]: E1001 16:35:06.383670 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:35:20 crc kubenswrapper[4688]: I1001 16:35:20.382890 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:35:20 crc kubenswrapper[4688]: E1001 16:35:20.383679 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:35:35 crc kubenswrapper[4688]: I1001 16:35:35.381294 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:35:35 crc kubenswrapper[4688]: E1001 16:35:35.382208 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:35:50 crc kubenswrapper[4688]: I1001 16:35:50.380996 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:35:50 crc kubenswrapper[4688]: E1001 16:35:50.381848 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.016457 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017338 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017384 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017419 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="extract-content" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017428 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="extract-content" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017441 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017449 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017473 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149aba72-0ea1-4a18-9ff1-98cfd0ada453" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017483 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="149aba72-0ea1-4a18-9ff1-98cfd0ada453" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017510 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="extract-content" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017555 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="extract-content" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017581 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="extract-utilities" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017590 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="extract-utilities" Oct 01 16:35:51 crc kubenswrapper[4688]: E1001 16:35:51.017605 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="extract-utilities" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.017613 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="extract-utilities" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.018004 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a303869d-96d7-41f0-ab91-37b2505fd625" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.018028 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="149aba72-0ea1-4a18-9ff1-98cfd0ada453" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.018057 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="773b3137-2c72-460f-9505-eb2a45c107fc" containerName="registry-server" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.019120 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.021990 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.022432 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gfb9h" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.022756 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.023032 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.036110 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.126484 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.126861 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127022 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127118 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127177 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127275 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t84mk\" (UniqueName: \"kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127325 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127492 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.127705 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229143 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229210 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229235 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229274 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229295 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229333 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t84mk\" (UniqueName: \"kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229357 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229393 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229445 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.229754 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.230292 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.230694 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.231077 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.231553 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.235361 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.236786 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.238412 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.249048 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t84mk\" (UniqueName: \"kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.260150 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.344985 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:35:51 crc kubenswrapper[4688]: I1001 16:35:51.817180 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 16:35:51 crc kubenswrapper[4688]: W1001 16:35:51.830164 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52eda021_1010_4c20_8a69_9b7df698f765.slice/crio-d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d WatchSource:0}: Error finding container d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d: Status 404 returned error can't find the container with id d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d Oct 01 16:35:52 crc kubenswrapper[4688]: I1001 16:35:52.828208 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"52eda021-1010-4c20-8a69-9b7df698f765","Type":"ContainerStarted","Data":"d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d"} Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.839013 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.841920 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.871325 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.945697 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9kjm\" (UniqueName: \"kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.945888 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:55 crc kubenswrapper[4688]: I1001 16:35:55.946659 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.047970 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9kjm\" (UniqueName: \"kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.048132 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.048166 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.048872 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.048895 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.077585 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9kjm\" (UniqueName: \"kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm\") pod \"community-operators-tdhvv\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:56 crc kubenswrapper[4688]: I1001 16:35:56.178201 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:35:57 crc kubenswrapper[4688]: I1001 16:35:57.850239 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:35:57 crc kubenswrapper[4688]: I1001 16:35:57.907940 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerStarted","Data":"7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5"} Oct 01 16:35:58 crc kubenswrapper[4688]: I1001 16:35:58.919747 4688 generic.go:334] "Generic (PLEG): container finished" podID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerID="a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322" exitCode=0 Oct 01 16:35:58 crc kubenswrapper[4688]: I1001 16:35:58.919816 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerDied","Data":"a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322"} Oct 01 16:35:59 crc kubenswrapper[4688]: I1001 16:35:59.934176 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerStarted","Data":"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8"} Oct 01 16:36:01 crc kubenswrapper[4688]: I1001 16:36:01.387677 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:36:01 crc kubenswrapper[4688]: E1001 16:36:01.388184 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:36:02 crc kubenswrapper[4688]: I1001 16:36:02.965034 4688 generic.go:334] "Generic (PLEG): container finished" podID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerID="281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8" exitCode=0 Oct 01 16:36:02 crc kubenswrapper[4688]: I1001 16:36:02.965099 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerDied","Data":"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8"} Oct 01 16:36:12 crc kubenswrapper[4688]: I1001 16:36:12.381108 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:36:12 crc kubenswrapper[4688]: E1001 16:36:12.381881 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:36:24 crc kubenswrapper[4688]: I1001 16:36:24.386646 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:36:24 crc kubenswrapper[4688]: E1001 16:36:24.387436 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:36:34 crc kubenswrapper[4688]: I1001 16:36:34.906328 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:36:34 crc kubenswrapper[4688]: E1001 16:36:34.960444 4688 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 01 16:36:34 crc kubenswrapper[4688]: E1001 16:36:34.961714 4688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t84mk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(52eda021-1010-4c20-8a69-9b7df698f765): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 16:36:34 crc kubenswrapper[4688]: E1001 16:36:34.962950 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="52eda021-1010-4c20-8a69-9b7df698f765" Oct 01 16:36:35 crc kubenswrapper[4688]: E1001 16:36:35.341778 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="52eda021-1010-4c20-8a69-9b7df698f765" Oct 01 16:36:36 crc kubenswrapper[4688]: I1001 16:36:36.349030 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerStarted","Data":"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d"} Oct 01 16:36:36 crc kubenswrapper[4688]: I1001 16:36:36.370236 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdhvv" podStartSLOduration=4.888987717 podStartE2EDuration="41.37021069s" podCreationTimestamp="2025-10-01 16:35:55 +0000 UTC" firstStartedPulling="2025-10-01 16:35:58.923627844 +0000 UTC m=+2948.274267816" lastFinishedPulling="2025-10-01 16:36:35.404850787 +0000 UTC m=+2984.755490789" observedRunningTime="2025-10-01 16:36:36.363325324 +0000 UTC m=+2985.713965286" watchObservedRunningTime="2025-10-01 16:36:36.37021069 +0000 UTC m=+2985.720850662" Oct 01 16:36:36 crc kubenswrapper[4688]: I1001 16:36:36.380761 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:36:36 crc kubenswrapper[4688]: E1001 16:36:36.381015 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.596668 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.598921 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.618416 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.672177 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.672270 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.672366 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42jmw\" (UniqueName: \"kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.774378 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.774443 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.774509 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42jmw\" (UniqueName: \"kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.775020 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.775210 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.799350 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42jmw\" (UniqueName: \"kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw\") pod \"redhat-operators-dv8z6\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:44 crc kubenswrapper[4688]: I1001 16:36:44.972002 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:45 crc kubenswrapper[4688]: I1001 16:36:45.484434 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.179327 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.179750 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.241476 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.468911 4688 generic.go:334] "Generic (PLEG): container finished" podID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerID="da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6" exitCode=0 Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.470669 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerDied","Data":"da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6"} Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.470738 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerStarted","Data":"82e08cace7eaf6c023722e274877033e0f044ed57133a3a4bfa5e38b259f86e8"} Oct 01 16:36:46 crc kubenswrapper[4688]: I1001 16:36:46.535606 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:47 crc kubenswrapper[4688]: I1001 16:36:47.381309 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:36:47 crc kubenswrapper[4688]: E1001 16:36:47.381560 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:36:47 crc kubenswrapper[4688]: I1001 16:36:47.482699 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerStarted","Data":"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d"} Oct 01 16:36:50 crc kubenswrapper[4688]: I1001 16:36:50.773949 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:36:50 crc kubenswrapper[4688]: I1001 16:36:50.774747 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdhvv" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="registry-server" containerID="cri-o://bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d" gracePeriod=2 Oct 01 16:36:50 crc kubenswrapper[4688]: I1001 16:36:50.956999 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.263878 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.323422 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content\") pod \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.323743 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities\") pod \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.323799 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9kjm\" (UniqueName: \"kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm\") pod \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\" (UID: \"3094cc42-212f-4fcb-83bf-f2623f4b04f1\") " Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.325385 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities" (OuterVolumeSpecName: "utilities") pod "3094cc42-212f-4fcb-83bf-f2623f4b04f1" (UID: "3094cc42-212f-4fcb-83bf-f2623f4b04f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.330483 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm" (OuterVolumeSpecName: "kube-api-access-f9kjm") pod "3094cc42-212f-4fcb-83bf-f2623f4b04f1" (UID: "3094cc42-212f-4fcb-83bf-f2623f4b04f1"). InnerVolumeSpecName "kube-api-access-f9kjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.359753 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3094cc42-212f-4fcb-83bf-f2623f4b04f1" (UID: "3094cc42-212f-4fcb-83bf-f2623f4b04f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.427159 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9kjm\" (UniqueName: \"kubernetes.io/projected/3094cc42-212f-4fcb-83bf-f2623f4b04f1-kube-api-access-f9kjm\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.427196 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.427209 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3094cc42-212f-4fcb-83bf-f2623f4b04f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.526347 4688 generic.go:334] "Generic (PLEG): container finished" podID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerID="5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d" exitCode=0 Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.526407 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerDied","Data":"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d"} Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.532163 4688 generic.go:334] "Generic (PLEG): container finished" podID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerID="bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d" exitCode=0 Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.532199 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerDied","Data":"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d"} Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.532222 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdhvv" event={"ID":"3094cc42-212f-4fcb-83bf-f2623f4b04f1","Type":"ContainerDied","Data":"7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5"} Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.532237 4688 scope.go:117] "RemoveContainer" containerID="bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.532361 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdhvv" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.573188 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.576029 4688 scope.go:117] "RemoveContainer" containerID="281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.582654 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdhvv"] Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.603674 4688 scope.go:117] "RemoveContainer" containerID="a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.624802 4688 scope.go:117] "RemoveContainer" containerID="bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d" Oct 01 16:36:51 crc kubenswrapper[4688]: E1001 16:36:51.625418 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d\": container with ID starting with bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d not found: ID does not exist" containerID="bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.625470 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d"} err="failed to get container status \"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d\": rpc error: code = NotFound desc = could not find container \"bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d\": container with ID starting with bd5d2c9f61f0a7b1bd19a51ad01028f26af026c0b5742e3bcea70fdf217aa54d not found: ID does not exist" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.625503 4688 scope.go:117] "RemoveContainer" containerID="281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8" Oct 01 16:36:51 crc kubenswrapper[4688]: E1001 16:36:51.625932 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8\": container with ID starting with 281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8 not found: ID does not exist" containerID="281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.625957 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8"} err="failed to get container status \"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8\": rpc error: code = NotFound desc = could not find container \"281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8\": container with ID starting with 281141f1a63aab1ee9a0af6cb5e04161173a7f9d6aa075de2f5da01ac279a2e8 not found: ID does not exist" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.625972 4688 scope.go:117] "RemoveContainer" containerID="a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322" Oct 01 16:36:51 crc kubenswrapper[4688]: E1001 16:36:51.626667 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322\": container with ID starting with a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322 not found: ID does not exist" containerID="a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322" Oct 01 16:36:51 crc kubenswrapper[4688]: I1001 16:36:51.626703 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322"} err="failed to get container status \"a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322\": rpc error: code = NotFound desc = could not find container \"a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322\": container with ID starting with a9ec8a6f4861e4bb93c7a4f7fc72d0cb0582eb2182987fa4127323bd0d2aa322 not found: ID does not exist" Oct 01 16:36:52 crc kubenswrapper[4688]: I1001 16:36:52.543997 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"52eda021-1010-4c20-8a69-9b7df698f765","Type":"ContainerStarted","Data":"126ba995c026e534bcacad6bb78f295801a0758c1210df0ee08c91e695f5d500"} Oct 01 16:36:52 crc kubenswrapper[4688]: I1001 16:36:52.546303 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerStarted","Data":"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5"} Oct 01 16:36:52 crc kubenswrapper[4688]: I1001 16:36:52.569309 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.453113326 podStartE2EDuration="1m3.569292475s" podCreationTimestamp="2025-10-01 16:35:49 +0000 UTC" firstStartedPulling="2025-10-01 16:35:51.835064344 +0000 UTC m=+2941.185704316" lastFinishedPulling="2025-10-01 16:36:50.951243493 +0000 UTC m=+3000.301883465" observedRunningTime="2025-10-01 16:36:52.566977699 +0000 UTC m=+3001.917617661" watchObservedRunningTime="2025-10-01 16:36:52.569292475 +0000 UTC m=+3001.919932437" Oct 01 16:36:52 crc kubenswrapper[4688]: I1001 16:36:52.596014 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dv8z6" podStartSLOduration=2.887762007 podStartE2EDuration="8.595994546s" podCreationTimestamp="2025-10-01 16:36:44 +0000 UTC" firstStartedPulling="2025-10-01 16:36:46.471930224 +0000 UTC m=+2995.822570196" lastFinishedPulling="2025-10-01 16:36:52.180162773 +0000 UTC m=+3001.530802735" observedRunningTime="2025-10-01 16:36:52.590591722 +0000 UTC m=+3001.941231694" watchObservedRunningTime="2025-10-01 16:36:52.595994546 +0000 UTC m=+3001.946634518" Oct 01 16:36:53 crc kubenswrapper[4688]: I1001 16:36:53.396214 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" path="/var/lib/kubelet/pods/3094cc42-212f-4fcb-83bf-f2623f4b04f1/volumes" Oct 01 16:36:54 crc kubenswrapper[4688]: I1001 16:36:54.972763 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:54 crc kubenswrapper[4688]: I1001 16:36:54.973242 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:36:56 crc kubenswrapper[4688]: I1001 16:36:56.033086 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dv8z6" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="registry-server" probeResult="failure" output=< Oct 01 16:36:56 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:36:56 crc kubenswrapper[4688]: > Oct 01 16:36:59 crc kubenswrapper[4688]: E1001 16:36:59.129391 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:02 crc kubenswrapper[4688]: I1001 16:37:02.381788 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:37:02 crc kubenswrapper[4688]: E1001 16:37:02.382600 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:37:05 crc kubenswrapper[4688]: I1001 16:37:05.030678 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:37:05 crc kubenswrapper[4688]: I1001 16:37:05.091652 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:37:05 crc kubenswrapper[4688]: I1001 16:37:05.270112 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:37:06 crc kubenswrapper[4688]: I1001 16:37:06.668117 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dv8z6" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="registry-server" containerID="cri-o://9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5" gracePeriod=2 Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.175493 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.341109 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content\") pod \"61b0b6a8-757a-42bf-9913-ad764780f07d\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.341151 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42jmw\" (UniqueName: \"kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw\") pod \"61b0b6a8-757a-42bf-9913-ad764780f07d\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.341260 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities\") pod \"61b0b6a8-757a-42bf-9913-ad764780f07d\" (UID: \"61b0b6a8-757a-42bf-9913-ad764780f07d\") " Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.342096 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities" (OuterVolumeSpecName: "utilities") pod "61b0b6a8-757a-42bf-9913-ad764780f07d" (UID: "61b0b6a8-757a-42bf-9913-ad764780f07d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.347203 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw" (OuterVolumeSpecName: "kube-api-access-42jmw") pod "61b0b6a8-757a-42bf-9913-ad764780f07d" (UID: "61b0b6a8-757a-42bf-9913-ad764780f07d"). InnerVolumeSpecName "kube-api-access-42jmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.445065 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42jmw\" (UniqueName: \"kubernetes.io/projected/61b0b6a8-757a-42bf-9913-ad764780f07d-kube-api-access-42jmw\") on node \"crc\" DevicePath \"\"" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.445112 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.455876 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61b0b6a8-757a-42bf-9913-ad764780f07d" (UID: "61b0b6a8-757a-42bf-9913-ad764780f07d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.549094 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61b0b6a8-757a-42bf-9913-ad764780f07d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.677107 4688 generic.go:334] "Generic (PLEG): container finished" podID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerID="9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5" exitCode=0 Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.677148 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerDied","Data":"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5"} Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.677174 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dv8z6" event={"ID":"61b0b6a8-757a-42bf-9913-ad764780f07d","Type":"ContainerDied","Data":"82e08cace7eaf6c023722e274877033e0f044ed57133a3a4bfa5e38b259f86e8"} Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.677189 4688 scope.go:117] "RemoveContainer" containerID="9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.677310 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dv8z6" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.755788 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.755802 4688 scope.go:117] "RemoveContainer" containerID="5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.761351 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dv8z6"] Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.817680 4688 scope.go:117] "RemoveContainer" containerID="da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.863358 4688 scope.go:117] "RemoveContainer" containerID="9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5" Oct 01 16:37:07 crc kubenswrapper[4688]: E1001 16:37:07.863709 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5\": container with ID starting with 9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5 not found: ID does not exist" containerID="9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.863748 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5"} err="failed to get container status \"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5\": rpc error: code = NotFound desc = could not find container \"9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5\": container with ID starting with 9b72b032e411f89614a80d051e1ac6ac6dcc9142a9a96e540cb35529a9b4a5f5 not found: ID does not exist" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.863772 4688 scope.go:117] "RemoveContainer" containerID="5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d" Oct 01 16:37:07 crc kubenswrapper[4688]: E1001 16:37:07.864158 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d\": container with ID starting with 5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d not found: ID does not exist" containerID="5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.864274 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d"} err="failed to get container status \"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d\": rpc error: code = NotFound desc = could not find container \"5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d\": container with ID starting with 5765334cf7cde7c1b6ef3ebced9e5ac4b86cc09ddc28556c4b8994f86646711d not found: ID does not exist" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.864360 4688 scope.go:117] "RemoveContainer" containerID="da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6" Oct 01 16:37:07 crc kubenswrapper[4688]: E1001 16:37:07.864651 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6\": container with ID starting with da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6 not found: ID does not exist" containerID="da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6" Oct 01 16:37:07 crc kubenswrapper[4688]: I1001 16:37:07.864753 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6"} err="failed to get container status \"da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6\": rpc error: code = NotFound desc = could not find container \"da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6\": container with ID starting with da864ea1b07e43a4fab8bd2f9ab507c8d0ee61cd9199f60bca2f7097d81e99a6 not found: ID does not exist" Oct 01 16:37:09 crc kubenswrapper[4688]: E1001 16:37:09.384651 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:09 crc kubenswrapper[4688]: I1001 16:37:09.397083 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" path="/var/lib/kubelet/pods/61b0b6a8-757a-42bf-9913-ad764780f07d/volumes" Oct 01 16:37:15 crc kubenswrapper[4688]: I1001 16:37:15.381356 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:37:15 crc kubenswrapper[4688]: E1001 16:37:15.382247 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:37:19 crc kubenswrapper[4688]: E1001 16:37:19.630561 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:28 crc kubenswrapper[4688]: I1001 16:37:28.381176 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:37:28 crc kubenswrapper[4688]: E1001 16:37:28.382014 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:37:29 crc kubenswrapper[4688]: E1001 16:37:29.850684 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:39 crc kubenswrapper[4688]: I1001 16:37:39.382049 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:37:39 crc kubenswrapper[4688]: E1001 16:37:39.383467 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:37:40 crc kubenswrapper[4688]: E1001 16:37:40.115151 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:50 crc kubenswrapper[4688]: E1001 16:37:50.354649 4688 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3094cc42_212f_4fcb_83bf_f2623f4b04f1.slice/crio-7e32393c1e6cab7a2b8298ae142abcc8f4c707e96e8184bf9efe52e5db2cc7b5\": RecentStats: unable to find data in memory cache]" Oct 01 16:37:51 crc kubenswrapper[4688]: E1001 16:37:51.438957 4688 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/7c7e752629ddb6b3af925112cf2b10fc1df4121d937055eed1459965ac26af6b/diff" to get inode usage: stat /var/lib/containers/storage/overlay/7c7e752629ddb6b3af925112cf2b10fc1df4121d937055eed1459965ac26af6b/diff: no such file or directory, extraDiskErr: Oct 01 16:37:52 crc kubenswrapper[4688]: I1001 16:37:52.381211 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:37:52 crc kubenswrapper[4688]: E1001 16:37:52.381724 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:38:04 crc kubenswrapper[4688]: I1001 16:38:04.381305 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:38:04 crc kubenswrapper[4688]: E1001 16:38:04.382091 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:38:15 crc kubenswrapper[4688]: I1001 16:38:15.381817 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:38:15 crc kubenswrapper[4688]: E1001 16:38:15.384652 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:38:26 crc kubenswrapper[4688]: I1001 16:38:26.382056 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:38:26 crc kubenswrapper[4688]: E1001 16:38:26.382847 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:38:40 crc kubenswrapper[4688]: I1001 16:38:40.380493 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:38:40 crc kubenswrapper[4688]: E1001 16:38:40.381637 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:38:51 crc kubenswrapper[4688]: I1001 16:38:51.387797 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:38:51 crc kubenswrapper[4688]: E1001 16:38:51.390158 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:39:05 crc kubenswrapper[4688]: I1001 16:39:05.381734 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:39:05 crc kubenswrapper[4688]: E1001 16:39:05.382824 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:39:20 crc kubenswrapper[4688]: I1001 16:39:20.381705 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:39:20 crc kubenswrapper[4688]: E1001 16:39:20.383496 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:39:22 crc kubenswrapper[4688]: I1001 16:39:22.016577 4688 generic.go:334] "Generic (PLEG): container finished" podID="52eda021-1010-4c20-8a69-9b7df698f765" containerID="126ba995c026e534bcacad6bb78f295801a0758c1210df0ee08c91e695f5d500" exitCode=0 Oct 01 16:39:22 crc kubenswrapper[4688]: I1001 16:39:22.016703 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"52eda021-1010-4c20-8a69-9b7df698f765","Type":"ContainerDied","Data":"126ba995c026e534bcacad6bb78f295801a0758c1210df0ee08c91e695f5d500"} Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.531317 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611674 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611730 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611765 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611803 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611835 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611873 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.611976 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t84mk\" (UniqueName: \"kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.612020 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.612041 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data\") pod \"52eda021-1010-4c20-8a69-9b7df698f765\" (UID: \"52eda021-1010-4c20-8a69-9b7df698f765\") " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.612970 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data" (OuterVolumeSpecName: "config-data") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.613399 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.617369 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.617824 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.620800 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk" (OuterVolumeSpecName: "kube-api-access-t84mk") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "kube-api-access-t84mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.644243 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.649757 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.651679 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.683748 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "52eda021-1010-4c20-8a69-9b7df698f765" (UID: "52eda021-1010-4c20-8a69-9b7df698f765"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.714905 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t84mk\" (UniqueName: \"kubernetes.io/projected/52eda021-1010-4c20-8a69-9b7df698f765-kube-api-access-t84mk\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.714949 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.714961 4688 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52eda021-1010-4c20-8a69-9b7df698f765-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.714970 4688 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.714979 4688 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.715002 4688 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.715011 4688 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/52eda021-1010-4c20-8a69-9b7df698f765-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.715042 4688 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.715051 4688 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/52eda021-1010-4c20-8a69-9b7df698f765-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.739059 4688 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 01 16:39:23 crc kubenswrapper[4688]: I1001 16:39:23.816339 4688 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 01 16:39:24 crc kubenswrapper[4688]: I1001 16:39:24.039386 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"52eda021-1010-4c20-8a69-9b7df698f765","Type":"ContainerDied","Data":"d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d"} Oct 01 16:39:24 crc kubenswrapper[4688]: I1001 16:39:24.039426 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d375ff73be06e456ef4552f357bdfe315b4a9c5a8466164d13eeedc2b646606d" Oct 01 16:39:24 crc kubenswrapper[4688]: I1001 16:39:24.039440 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.736516 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741589 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52eda021-1010-4c20-8a69-9b7df698f765" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741621 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="52eda021-1010-4c20-8a69-9b7df698f765" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741655 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741661 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741678 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="extract-utilities" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741685 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="extract-utilities" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741704 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741710 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741724 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="extract-content" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741729 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="extract-content" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741740 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="extract-content" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741746 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="extract-content" Oct 01 16:39:27 crc kubenswrapper[4688]: E1001 16:39:27.741758 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="extract-utilities" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741764 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="extract-utilities" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741949 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="3094cc42-212f-4fcb-83bf-f2623f4b04f1" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741961 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b0b6a8-757a-42bf-9913-ad764780f07d" containerName="registry-server" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.741974 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="52eda021-1010-4c20-8a69-9b7df698f765" containerName="tempest-tests-tempest-tests-runner" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.742590 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.745017 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.747196 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gfb9h" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.801484 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.801666 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dnpx\" (UniqueName: \"kubernetes.io/projected/76b03e8a-66be-4233-99bd-3109365d180a-kube-api-access-8dnpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.903054 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dnpx\" (UniqueName: \"kubernetes.io/projected/76b03e8a-66be-4233-99bd-3109365d180a-kube-api-access-8dnpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.903168 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.903709 4688 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.933584 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dnpx\" (UniqueName: \"kubernetes.io/projected/76b03e8a-66be-4233-99bd-3109365d180a-kube-api-access-8dnpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:27 crc kubenswrapper[4688]: I1001 16:39:27.936363 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"76b03e8a-66be-4233-99bd-3109365d180a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:28 crc kubenswrapper[4688]: I1001 16:39:28.071437 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 16:39:28 crc kubenswrapper[4688]: I1001 16:39:28.516673 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 16:39:29 crc kubenswrapper[4688]: I1001 16:39:29.095307 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"76b03e8a-66be-4233-99bd-3109365d180a","Type":"ContainerStarted","Data":"ac210793d62fce090d4cc4580d03ad90184baa5fee614948c1621717a3520d50"} Oct 01 16:39:30 crc kubenswrapper[4688]: I1001 16:39:30.107861 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"76b03e8a-66be-4233-99bd-3109365d180a","Type":"ContainerStarted","Data":"dc689c2f77a9ab16f6bce50dbfa20aa33ed6c71a15f7f26289cf23ddfd4af7b5"} Oct 01 16:39:30 crc kubenswrapper[4688]: I1001 16:39:30.130476 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9592227979999999 podStartE2EDuration="3.130457434s" podCreationTimestamp="2025-10-01 16:39:27 +0000 UTC" firstStartedPulling="2025-10-01 16:39:28.529376346 +0000 UTC m=+3157.880016298" lastFinishedPulling="2025-10-01 16:39:29.700610972 +0000 UTC m=+3159.051250934" observedRunningTime="2025-10-01 16:39:30.125786541 +0000 UTC m=+3159.476426503" watchObservedRunningTime="2025-10-01 16:39:30.130457434 +0000 UTC m=+3159.481097396" Oct 01 16:39:34 crc kubenswrapper[4688]: I1001 16:39:34.382298 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:39:34 crc kubenswrapper[4688]: E1001 16:39:34.384949 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:39:46 crc kubenswrapper[4688]: I1001 16:39:46.381502 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:39:46 crc kubenswrapper[4688]: E1001 16:39:46.382790 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.399311 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wzjvs/must-gather-9c2wh"] Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.401077 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.406565 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wzjvs"/"openshift-service-ca.crt" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.407909 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wzjvs"/"kube-root-ca.crt" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.410769 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wzjvs"/"default-dockercfg-scls8" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.428254 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wzjvs/must-gather-9c2wh"] Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.433552 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg5ts\" (UniqueName: \"kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.454330 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.555841 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.555947 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg5ts\" (UniqueName: \"kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.556275 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.579496 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg5ts\" (UniqueName: \"kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts\") pod \"must-gather-9c2wh\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:47 crc kubenswrapper[4688]: I1001 16:39:47.720205 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:39:48 crc kubenswrapper[4688]: I1001 16:39:48.243632 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wzjvs/must-gather-9c2wh"] Oct 01 16:39:48 crc kubenswrapper[4688]: I1001 16:39:48.308602 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" event={"ID":"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2","Type":"ContainerStarted","Data":"c2e5c0d6528b7e751bdc6ce45e874d239164f426d4ef30efa1d0b0e04b94d582"} Oct 01 16:39:57 crc kubenswrapper[4688]: I1001 16:39:57.419585 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" event={"ID":"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2","Type":"ContainerStarted","Data":"56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6"} Oct 01 16:39:57 crc kubenswrapper[4688]: I1001 16:39:57.420152 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" event={"ID":"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2","Type":"ContainerStarted","Data":"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35"} Oct 01 16:39:57 crc kubenswrapper[4688]: I1001 16:39:57.446697 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" podStartSLOduration=1.843500734 podStartE2EDuration="10.44667035s" podCreationTimestamp="2025-10-01 16:39:47 +0000 UTC" firstStartedPulling="2025-10-01 16:39:48.25870826 +0000 UTC m=+3177.609348222" lastFinishedPulling="2025-10-01 16:39:56.861877876 +0000 UTC m=+3186.212517838" observedRunningTime="2025-10-01 16:39:57.436751208 +0000 UTC m=+3186.787391170" watchObservedRunningTime="2025-10-01 16:39:57.44667035 +0000 UTC m=+3186.797310312" Oct 01 16:40:00 crc kubenswrapper[4688]: I1001 16:40:00.380692 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.484253 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a"} Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.724893 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mxnwq"] Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.726197 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.847095 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9qfz\" (UniqueName: \"kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.847189 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.948495 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9qfz\" (UniqueName: \"kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.948886 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.948990 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:01 crc kubenswrapper[4688]: I1001 16:40:01.967986 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9qfz\" (UniqueName: \"kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz\") pod \"crc-debug-mxnwq\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:02 crc kubenswrapper[4688]: I1001 16:40:02.045090 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:40:02 crc kubenswrapper[4688]: I1001 16:40:02.495431 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" event={"ID":"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4","Type":"ContainerStarted","Data":"7f6a3aa715b6b3bba632a3764fbc70266380d3bc7990b28994e91bfbb4fffb06"} Oct 01 16:40:16 crc kubenswrapper[4688]: I1001 16:40:16.659159 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" event={"ID":"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4","Type":"ContainerStarted","Data":"d9e033b62daa6b5716fa17780c48cc1a30714760b9f6016a9b5f2b2dee355eb2"} Oct 01 16:40:16 crc kubenswrapper[4688]: I1001 16:40:16.684250 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" podStartSLOduration=2.044993859 podStartE2EDuration="15.68422262s" podCreationTimestamp="2025-10-01 16:40:01 +0000 UTC" firstStartedPulling="2025-10-01 16:40:02.078901015 +0000 UTC m=+3191.429540977" lastFinishedPulling="2025-10-01 16:40:15.718129776 +0000 UTC m=+3205.068769738" observedRunningTime="2025-10-01 16:40:16.670990723 +0000 UTC m=+3206.021630705" watchObservedRunningTime="2025-10-01 16:40:16.68422262 +0000 UTC m=+3206.034862592" Oct 01 16:41:12 crc kubenswrapper[4688]: I1001 16:41:12.402859 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8668d66c66-j8wfd_6de36fb1-5b60-4600-ae35-bd31fdec91ce/barbican-api-log/0.log" Oct 01 16:41:12 crc kubenswrapper[4688]: I1001 16:41:12.420710 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8668d66c66-j8wfd_6de36fb1-5b60-4600-ae35-bd31fdec91ce/barbican-api/0.log" Oct 01 16:41:12 crc kubenswrapper[4688]: I1001 16:41:12.948004 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-df57dbd74-8gcrt_d13b3026-f303-454c-a686-7bc22f5211cf/barbican-keystone-listener/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.197793 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-df57dbd74-8gcrt_d13b3026-f303-454c-a686-7bc22f5211cf/barbican-keystone-listener-log/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.241889 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74574df4f9-v4dnz_b07391ac-b11b-4b46-8e9b-847b424e7a80/barbican-worker/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.432949 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74574df4f9-v4dnz_b07391ac-b11b-4b46-8e9b-847b424e7a80/barbican-worker-log/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.512308 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bk825_fbd7ce66-5861-4a86-a892-ce6493baf848/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.765733 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/ceilometer-central-agent/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.810184 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/ceilometer-notification-agent/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.940896 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/proxy-httpd/0.log" Oct 01 16:41:13 crc kubenswrapper[4688]: I1001 16:41:13.969504 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/sg-core/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.152816 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_12a98d94-2c07-4fa8-b69b-c24a11fc6cac/cinder-api/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.190679 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_12a98d94-2c07-4fa8-b69b-c24a11fc6cac/cinder-api-log/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.421158 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_825e667e-354c-4cd3-bb07-d51c67981ac8/cinder-scheduler/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.461961 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_825e667e-354c-4cd3-bb07-d51c67981ac8/probe/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.741478 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k4lch_3fb9e457-e82c-4418-a5a6-a3892a359683/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:14 crc kubenswrapper[4688]: I1001 16:41:14.886511 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6_42f771c8-2902-4358-8f27-3dd4e9ac510e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.011392 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk_4633912c-f539-44fd-8fc4-1e57ddbdad56/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.253637 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/init/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.602354 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/dnsmasq-dns/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.629887 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/init/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.813052 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f_9cd2ef48-3ade-4edb-bd61-c3449a9c64fe/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:15 crc kubenswrapper[4688]: I1001 16:41:15.870011 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_32ec1a41-4380-4f29-a56d-dd2d722999d6/glance-httpd/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.012098 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_32ec1a41-4380-4f29-a56d-dd2d722999d6/glance-log/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.163561 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ab7d6e5f-ceab-400c-858e-2059dc492923/glance-httpd/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.299457 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ab7d6e5f-ceab-400c-858e-2059dc492923/glance-log/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.651061 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.664471 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon/1.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.948351 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x_944f4da8-c2bb-400d-a78f-caf11336d5ec/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:16 crc kubenswrapper[4688]: I1001 16:41:16.971418 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon-log/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.090737 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pdvcs_77d2978e-28f0-4591-a3fc-491ebfb6d1d9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.379804 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c522fa53-fdbd-4836-b127-a858c5001cc0/kube-state-metrics/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.643683 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x_959f38f8-3df7-422b-9c09-a6fa68d4b8f0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.661670 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-587df57d87-hlpwt_8a94d646-b2be-4ab1-89ff-da8aa30591f6/keystone-api/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.951800 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c946d4bf9-l8zds_5766bcc5-cf95-400d-a634-c2e0ad4fe3b9/neutron-httpd/0.log" Oct 01 16:41:17 crc kubenswrapper[4688]: I1001 16:41:17.990634 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c946d4bf9-l8zds_5766bcc5-cf95-400d-a634-c2e0ad4fe3b9/neutron-api/0.log" Oct 01 16:41:18 crc kubenswrapper[4688]: I1001 16:41:18.245521 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz_0ddaaf82-b5e6-4960-b79e-77ad2f913b93/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:18 crc kubenswrapper[4688]: I1001 16:41:18.786102 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cdb337c8-732c-4123-b47b-212ecb38257d/nova-api-log/0.log" Oct 01 16:41:18 crc kubenswrapper[4688]: I1001 16:41:18.990651 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cdb337c8-732c-4123-b47b-212ecb38257d/nova-api-api/0.log" Oct 01 16:41:19 crc kubenswrapper[4688]: I1001 16:41:19.118296 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f64f51ee-f2ae-4c33-babf-67838b0f0135/nova-cell0-conductor-conductor/0.log" Oct 01 16:41:19 crc kubenswrapper[4688]: I1001 16:41:19.406848 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2/nova-cell1-conductor-conductor/0.log" Oct 01 16:41:19 crc kubenswrapper[4688]: I1001 16:41:19.497662 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_050f05f1-0f70-46e1-94a3-e32d2aa42c36/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 16:41:19 crc kubenswrapper[4688]: I1001 16:41:19.795946 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8g64m_6206a3c7-1bb2-422a-8bc9-38e62d890513/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:20 crc kubenswrapper[4688]: I1001 16:41:20.182037 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c9f50e4d-f039-4902-ab15-03604f701daf/nova-metadata-log/0.log" Oct 01 16:41:20 crc kubenswrapper[4688]: I1001 16:41:20.563778 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8d577d2f-a1a6-4d70-8532-86c0bb41f51f/nova-scheduler-scheduler/0.log" Oct 01 16:41:20 crc kubenswrapper[4688]: I1001 16:41:20.786827 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/mysql-bootstrap/0.log" Oct 01 16:41:20 crc kubenswrapper[4688]: I1001 16:41:20.999781 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/mysql-bootstrap/0.log" Oct 01 16:41:21 crc kubenswrapper[4688]: I1001 16:41:21.069680 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/galera/0.log" Oct 01 16:41:21 crc kubenswrapper[4688]: I1001 16:41:21.171854 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c9f50e4d-f039-4902-ab15-03604f701daf/nova-metadata-metadata/0.log" Oct 01 16:41:21 crc kubenswrapper[4688]: I1001 16:41:21.444402 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/mysql-bootstrap/0.log" Oct 01 16:41:21 crc kubenswrapper[4688]: I1001 16:41:21.732507 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/galera/0.log" Oct 01 16:41:21 crc kubenswrapper[4688]: I1001 16:41:21.737685 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/mysql-bootstrap/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.016460 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a309f66f-cc9d-40e7-9acc-00881395f4c9/openstackclient/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.082481 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sfjc5_9b3cd14b-091c-461a-8ce5-924f9619a596/openstack-network-exporter/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.383184 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server-init/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.625563 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server-init/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.634319 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovs-vswitchd/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.657000 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server/0.log" Oct 01 16:41:22 crc kubenswrapper[4688]: I1001 16:41:22.918433 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ts722_e4559e8a-674a-427d-a9fb-145f94ff3b58/ovn-controller/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.101870 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-t4nl2_da404969-614a-4c71-9ce3-0a15a702aa65/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.233371 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_33e5a114-5ed4-4df4-88ca-908340de2204/openstack-network-exporter/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.343098 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_33e5a114-5ed4-4df4-88ca-908340de2204/ovn-northd/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.560959 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79a0889a-6692-43c9-928b-13e06ccd6ca1/openstack-network-exporter/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.598925 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79a0889a-6692-43c9-928b-13e06ccd6ca1/ovsdbserver-nb/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.821737 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_466a92b3-31f5-41fb-8810-76e95dd7a332/ovsdbserver-sb/0.log" Oct 01 16:41:23 crc kubenswrapper[4688]: I1001 16:41:23.910792 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_466a92b3-31f5-41fb-8810-76e95dd7a332/openstack-network-exporter/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.153359 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-76d4fdc768-gl782_d97f10ed-cfae-4409-b6fd-ae047f584ac3/placement-api/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.197945 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-76d4fdc768-gl782_d97f10ed-cfae-4409-b6fd-ae047f584ac3/placement-log/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.411694 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/setup-container/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.645545 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/setup-container/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.687631 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/rabbitmq/0.log" Oct 01 16:41:24 crc kubenswrapper[4688]: I1001 16:41:24.920221 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/setup-container/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.090424 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/setup-container/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.124964 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/rabbitmq/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.341042 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-c755p_9cebafe1-24f0-4751-b691-8fd0bc09359f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.486092 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-7tdxj_1d630ed1-0ec0-4b50-aef4-b8a75aa67270/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.671837 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc_d8469466-3ce0-4d29-8d96-6cce137617fa/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:25 crc kubenswrapper[4688]: I1001 16:41:25.910646 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-r7vc9_a1720cc4-8ce5-4d4a-b0f0-47afac314627/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.084569 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bj9n4_289bca66-8942-4580-b92e-d5cb8db5deed/ssh-known-hosts-edpm-deployment/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.416718 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bbc5fdc47-8258p_43929b9b-990a-4a85-9100-509d80bba3c1/proxy-server/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.512934 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bbc5fdc47-8258p_43929b9b-990a-4a85-9100-509d80bba3c1/proxy-httpd/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.715875 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-96jrl_f47bc7a7-66e7-4994-985d-2ff7da317e55/swift-ring-rebalance/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.794470 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-auditor/0.log" Oct 01 16:41:26 crc kubenswrapper[4688]: I1001 16:41:26.981986 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-reaper/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.071471 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-replicator/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.113163 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-server/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.274252 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-auditor/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.418928 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-server/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.419596 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-replicator/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.553339 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-updater/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.677136 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-auditor/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.712796 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-expirer/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.861437 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-replicator/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.932087 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-server/0.log" Oct 01 16:41:27 crc kubenswrapper[4688]: I1001 16:41:27.957628 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-updater/0.log" Oct 01 16:41:28 crc kubenswrapper[4688]: I1001 16:41:28.153456 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/rsync/0.log" Oct 01 16:41:28 crc kubenswrapper[4688]: I1001 16:41:28.174204 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/swift-recon-cron/0.log" Oct 01 16:41:28 crc kubenswrapper[4688]: I1001 16:41:28.532755 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2_149aba72-0ea1-4a18-9ff1-98cfd0ada453/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:28 crc kubenswrapper[4688]: I1001 16:41:28.596012 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_52eda021-1010-4c20-8a69-9b7df698f765/tempest-tests-tempest-tests-runner/0.log" Oct 01 16:41:28 crc kubenswrapper[4688]: I1001 16:41:28.788112 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_76b03e8a-66be-4233-99bd-3109365d180a/test-operator-logs-container/0.log" Oct 01 16:41:29 crc kubenswrapper[4688]: I1001 16:41:29.047249 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-whw8p_fd4cc96c-61fa-4f47-8a36-8d2c6ddef862/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:41:34 crc kubenswrapper[4688]: I1001 16:41:34.536705 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3720d335-ad5b-41cf-9acb-7f9b49fb2c6c/memcached/0.log" Oct 01 16:42:18 crc kubenswrapper[4688]: I1001 16:42:18.931388 4688 generic.go:334] "Generic (PLEG): container finished" podID="a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" containerID="d9e033b62daa6b5716fa17780c48cc1a30714760b9f6016a9b5f2b2dee355eb2" exitCode=0 Oct 01 16:42:18 crc kubenswrapper[4688]: I1001 16:42:18.931498 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" event={"ID":"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4","Type":"ContainerDied","Data":"d9e033b62daa6b5716fa17780c48cc1a30714760b9f6016a9b5f2b2dee355eb2"} Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.046336 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.087430 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mxnwq"] Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.097434 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mxnwq"] Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.119671 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host\") pod \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.119753 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9qfz\" (UniqueName: \"kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz\") pod \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\" (UID: \"a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4\") " Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.120749 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host" (OuterVolumeSpecName: "host") pod "a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" (UID: "a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.131808 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz" (OuterVolumeSpecName: "kube-api-access-w9qfz") pod "a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" (UID: "a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4"). InnerVolumeSpecName "kube-api-access-w9qfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.222364 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9qfz\" (UniqueName: \"kubernetes.io/projected/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-kube-api-access-w9qfz\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.222398 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.962083 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f6a3aa715b6b3bba632a3764fbc70266380d3bc7990b28994e91bfbb4fffb06" Oct 01 16:42:20 crc kubenswrapper[4688]: I1001 16:42:20.962362 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mxnwq" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.280745 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mhd9m"] Oct 01 16:42:21 crc kubenswrapper[4688]: E1001 16:42:21.281167 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" containerName="container-00" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.281181 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" containerName="container-00" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.281420 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" containerName="container-00" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.282102 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.342792 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpn2b\" (UniqueName: \"kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.342887 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.414406 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4" path="/var/lib/kubelet/pods/a99f2d1d-dbea-4510-9b92-b9bdd8fdd2b4/volumes" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.446069 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpn2b\" (UniqueName: \"kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.446327 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.446711 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.473196 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpn2b\" (UniqueName: \"kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b\") pod \"crc-debug-mhd9m\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.605619 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.975395 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" event={"ID":"86bddb2e-d30c-4dbb-81a8-89950a299080","Type":"ContainerStarted","Data":"51c65dbb5b42c526a743f2e9fa40c6ef3a3d3793fabaf8427a0d6ea93937232f"} Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.975828 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" event={"ID":"86bddb2e-d30c-4dbb-81a8-89950a299080","Type":"ContainerStarted","Data":"1cca6ecc4c2058a0543d1ba08ce3779674765ae87396fcd66f8e95b7e2e0d59a"} Oct 01 16:42:21 crc kubenswrapper[4688]: I1001 16:42:21.989241 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" podStartSLOduration=0.989221896 podStartE2EDuration="989.221896ms" podCreationTimestamp="2025-10-01 16:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:42:21.9848012 +0000 UTC m=+3331.335441162" watchObservedRunningTime="2025-10-01 16:42:21.989221896 +0000 UTC m=+3331.339861868" Oct 01 16:42:22 crc kubenswrapper[4688]: I1001 16:42:22.982748 4688 generic.go:334] "Generic (PLEG): container finished" podID="86bddb2e-d30c-4dbb-81a8-89950a299080" containerID="51c65dbb5b42c526a743f2e9fa40c6ef3a3d3793fabaf8427a0d6ea93937232f" exitCode=0 Oct 01 16:42:22 crc kubenswrapper[4688]: I1001 16:42:22.982796 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" event={"ID":"86bddb2e-d30c-4dbb-81a8-89950a299080","Type":"ContainerDied","Data":"51c65dbb5b42c526a743f2e9fa40c6ef3a3d3793fabaf8427a0d6ea93937232f"} Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.089161 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.194892 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host\") pod \"86bddb2e-d30c-4dbb-81a8-89950a299080\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.195137 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host" (OuterVolumeSpecName: "host") pod "86bddb2e-d30c-4dbb-81a8-89950a299080" (UID: "86bddb2e-d30c-4dbb-81a8-89950a299080"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.195644 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpn2b\" (UniqueName: \"kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b\") pod \"86bddb2e-d30c-4dbb-81a8-89950a299080\" (UID: \"86bddb2e-d30c-4dbb-81a8-89950a299080\") " Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.197936 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86bddb2e-d30c-4dbb-81a8-89950a299080-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.204390 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b" (OuterVolumeSpecName: "kube-api-access-xpn2b") pod "86bddb2e-d30c-4dbb-81a8-89950a299080" (UID: "86bddb2e-d30c-4dbb-81a8-89950a299080"). InnerVolumeSpecName "kube-api-access-xpn2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:42:24 crc kubenswrapper[4688]: I1001 16:42:24.299133 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpn2b\" (UniqueName: \"kubernetes.io/projected/86bddb2e-d30c-4dbb-81a8-89950a299080-kube-api-access-xpn2b\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:25 crc kubenswrapper[4688]: I1001 16:42:25.004837 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" event={"ID":"86bddb2e-d30c-4dbb-81a8-89950a299080","Type":"ContainerDied","Data":"1cca6ecc4c2058a0543d1ba08ce3779674765ae87396fcd66f8e95b7e2e0d59a"} Oct 01 16:42:25 crc kubenswrapper[4688]: I1001 16:42:25.004886 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cca6ecc4c2058a0543d1ba08ce3779674765ae87396fcd66f8e95b7e2e0d59a" Oct 01 16:42:25 crc kubenswrapper[4688]: I1001 16:42:25.004889 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-mhd9m" Oct 01 16:42:25 crc kubenswrapper[4688]: I1001 16:42:25.052620 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:42:25 crc kubenswrapper[4688]: I1001 16:42:25.052670 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:42:28 crc kubenswrapper[4688]: I1001 16:42:28.041570 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mhd9m"] Oct 01 16:42:28 crc kubenswrapper[4688]: I1001 16:42:28.055181 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-mhd9m"] Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.278115 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-dw4d8"] Oct 01 16:42:29 crc kubenswrapper[4688]: E1001 16:42:29.278928 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bddb2e-d30c-4dbb-81a8-89950a299080" containerName="container-00" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.278961 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bddb2e-d30c-4dbb-81a8-89950a299080" containerName="container-00" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.279374 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="86bddb2e-d30c-4dbb-81a8-89950a299080" containerName="container-00" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.280735 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.381967 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.382383 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9hb\" (UniqueName: \"kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.396614 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86bddb2e-d30c-4dbb-81a8-89950a299080" path="/var/lib/kubelet/pods/86bddb2e-d30c-4dbb-81a8-89950a299080/volumes" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.484499 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9hb\" (UniqueName: \"kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.484801 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.485411 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.514181 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9hb\" (UniqueName: \"kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb\") pod \"crc-debug-dw4d8\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:29 crc kubenswrapper[4688]: I1001 16:42:29.597406 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:30 crc kubenswrapper[4688]: I1001 16:42:30.054901 4688 generic.go:334] "Generic (PLEG): container finished" podID="c0ece00f-3a01-48a5-9a3a-1f34b7834234" containerID="4e3e1e101062be354af95ead8905194fa473d37000d75266b03c28a1a453538b" exitCode=0 Oct 01 16:42:30 crc kubenswrapper[4688]: I1001 16:42:30.054992 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" event={"ID":"c0ece00f-3a01-48a5-9a3a-1f34b7834234","Type":"ContainerDied","Data":"4e3e1e101062be354af95ead8905194fa473d37000d75266b03c28a1a453538b"} Oct 01 16:42:30 crc kubenswrapper[4688]: I1001 16:42:30.055269 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" event={"ID":"c0ece00f-3a01-48a5-9a3a-1f34b7834234","Type":"ContainerStarted","Data":"125f91baf6837c7868b566c0beab5e0d6cfe9935f536c44b660f9ea9c06ee308"} Oct 01 16:42:30 crc kubenswrapper[4688]: I1001 16:42:30.110090 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-dw4d8"] Oct 01 16:42:30 crc kubenswrapper[4688]: I1001 16:42:30.121283 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wzjvs/crc-debug-dw4d8"] Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.168023 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.215420 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q9hb\" (UniqueName: \"kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb\") pod \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.215562 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host\") pod \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\" (UID: \"c0ece00f-3a01-48a5-9a3a-1f34b7834234\") " Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.215868 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host" (OuterVolumeSpecName: "host") pod "c0ece00f-3a01-48a5-9a3a-1f34b7834234" (UID: "c0ece00f-3a01-48a5-9a3a-1f34b7834234"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.216199 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c0ece00f-3a01-48a5-9a3a-1f34b7834234-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.220814 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb" (OuterVolumeSpecName: "kube-api-access-6q9hb") pod "c0ece00f-3a01-48a5-9a3a-1f34b7834234" (UID: "c0ece00f-3a01-48a5-9a3a-1f34b7834234"). InnerVolumeSpecName "kube-api-access-6q9hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.318394 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q9hb\" (UniqueName: \"kubernetes.io/projected/c0ece00f-3a01-48a5-9a3a-1f34b7834234-kube-api-access-6q9hb\") on node \"crc\" DevicePath \"\"" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.392335 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ece00f-3a01-48a5-9a3a-1f34b7834234" path="/var/lib/kubelet/pods/c0ece00f-3a01-48a5-9a3a-1f34b7834234/volumes" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.762211 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.925403 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.953887 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:42:31 crc kubenswrapper[4688]: I1001 16:42:31.987876 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.072417 4688 scope.go:117] "RemoveContainer" containerID="4e3e1e101062be354af95ead8905194fa473d37000d75266b03c28a1a453538b" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.072446 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/crc-debug-dw4d8" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.164717 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.164942 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.223797 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/extract/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.400942 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-65bw4_06f11b98-29b6-41f9-a9fa-771b2f9565ad/kube-rbac-proxy/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.424707 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-65bw4_06f11b98-29b6-41f9-a9fa-771b2f9565ad/manager/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.562819 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gbg2c_5c1fc770-71c1-4e90-b803-d0d020cf22dc/kube-rbac-proxy/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.633341 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gbg2c_5c1fc770-71c1-4e90-b803-d0d020cf22dc/manager/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.653087 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-8pnfq_5fce237a-66be-49ef-90a8-cd0913daf883/kube-rbac-proxy/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.773909 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-8pnfq_5fce237a-66be-49ef-90a8-cd0913daf883/manager/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.852358 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-548l2_09bbe1ee-12b6-4342-b585-f5dfebdcfb79/kube-rbac-proxy/0.log" Oct 01 16:42:32 crc kubenswrapper[4688]: I1001 16:42:32.950216 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-548l2_09bbe1ee-12b6-4342-b585-f5dfebdcfb79/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.111288 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-bh7tx_e983d733-09fb-4a40-902b-f180b58d5b7b/kube-rbac-proxy/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.124587 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-bh7tx_e983d733-09fb-4a40-902b-f180b58d5b7b/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.322003 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5qm6b_44ef9a6f-0918-43c7-b153-abd8c7635131/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.351452 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5qm6b_44ef9a6f-0918-43c7-b153-abd8c7635131/kube-rbac-proxy/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.456895 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-6ccdl_dc2fb11b-b04c-444c-8623-43a6b566cf6c/kube-rbac-proxy/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.677845 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-6ccdl_dc2fb11b-b04c-444c-8623-43a6b566cf6c/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.684430 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-b8dfc_2fbdd013-920f-4e46-8519-3f8bbe24f25d/kube-rbac-proxy/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.735797 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-b8dfc_2fbdd013-920f-4e46-8519-3f8bbe24f25d/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.930272 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-85595f4675-l62s8_57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c/manager/0.log" Oct 01 16:42:33 crc kubenswrapper[4688]: I1001 16:42:33.959435 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-85595f4675-l62s8_57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c/kube-rbac-proxy/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.324607 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-mbsl5_45f48342-2b0d-4376-87ac-0c775c863326/manager/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.493003 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-mbsl5_45f48342-2b0d-4376-87ac-0c775c863326/kube-rbac-proxy/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.548926 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-x8lrd_8fee523c-d09e-422d-94ae-b85585856d27/kube-rbac-proxy/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.577673 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-x8lrd_8fee523c-d09e-422d-94ae-b85585856d27/manager/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.779985 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-h7xkj_6cfc7f51-a252-4b73-954c-b9bb611bf88e/kube-rbac-proxy/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.843408 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-h7xkj_6cfc7f51-a252-4b73-954c-b9bb611bf88e/manager/0.log" Oct 01 16:42:34 crc kubenswrapper[4688]: I1001 16:42:34.994804 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-9ppnb_3c33683d-9791-4d98-8d6e-0e58f48b4d94/kube-rbac-proxy/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.097053 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-9ppnb_3c33683d-9791-4d98-8d6e-0e58f48b4d94/manager/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.133804 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-rvtpg_8c13457f-36f6-49b5-9ac2-95298f76cd07/kube-rbac-proxy/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.236001 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-rvtpg_8c13457f-36f6-49b5-9ac2-95298f76cd07/manager/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.348765 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cs62rh_eb295639-6a81-4c5c-a02b-4101bf064788/kube-rbac-proxy/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.402404 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cs62rh_eb295639-6a81-4c5c-a02b-4101bf064788/manager/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.435136 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c677785d7-6g6wn_7b66b48a-f1b3-4c89-a879-a662dce03612/kube-rbac-proxy/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.789229 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6445cc7789-t2zln_ded465e1-ef18-4321-9771-e3a3dd7a92e8/kube-rbac-proxy/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.907262 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9cs6j_20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010/registry-server/0.log" Oct 01 16:42:35 crc kubenswrapper[4688]: I1001 16:42:35.943169 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6445cc7789-t2zln_ded465e1-ef18-4321-9771-e3a3dd7a92e8/operator/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.026007 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-tp7l5_fc3e31c2-228a-4ba3-aba1-113ef34d4917/kube-rbac-proxy/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.293378 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g6725_7c87144a-f596-49d5-8344-b392d2bc4f51/kube-rbac-proxy/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.350287 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-tp7l5_fc3e31c2-228a-4ba3-aba1-113ef34d4917/manager/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.503066 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g6725_7c87144a-f596-49d5-8344-b392d2bc4f51/manager/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.578325 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c677785d7-6g6wn_7b66b48a-f1b3-4c89-a879-a662dce03612/manager/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.658468 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn_fa800656-3342-44ac-9ec5-537382b5c67e/operator/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.720293 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-f4v99_604860e3-3d42-47b7-9387-c54faa3c8ec7/kube-rbac-proxy/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.807485 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-f4v99_604860e3-3d42-47b7-9387-c54faa3c8ec7/manager/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.886904 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-44nm7_a9b72717-b2f8-4340-81ed-88d2c3e0f2d6/manager/0.log" Oct 01 16:42:36 crc kubenswrapper[4688]: I1001 16:42:36.905255 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-44nm7_a9b72717-b2f8-4340-81ed-88d2c3e0f2d6/kube-rbac-proxy/0.log" Oct 01 16:42:37 crc kubenswrapper[4688]: I1001 16:42:37.031125 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-7blrb_733cc8e0-867c-4c7a-b535-c2493a08ed9f/kube-rbac-proxy/0.log" Oct 01 16:42:37 crc kubenswrapper[4688]: I1001 16:42:37.117670 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-7blrb_733cc8e0-867c-4c7a-b535-c2493a08ed9f/manager/0.log" Oct 01 16:42:37 crc kubenswrapper[4688]: I1001 16:42:37.144408 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-27q6x_52bede76-5e78-4a87-b1ed-4ec996866099/kube-rbac-proxy/0.log" Oct 01 16:42:37 crc kubenswrapper[4688]: I1001 16:42:37.210245 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-27q6x_52bede76-5e78-4a87-b1ed-4ec996866099/manager/0.log" Oct 01 16:42:55 crc kubenswrapper[4688]: I1001 16:42:55.042143 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9kj7x_dc8ba611-f5be-48f7-990f-f56d4f7a9539/control-plane-machine-set-operator/0.log" Oct 01 16:42:55 crc kubenswrapper[4688]: I1001 16:42:55.053306 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:42:55 crc kubenswrapper[4688]: I1001 16:42:55.053366 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:42:55 crc kubenswrapper[4688]: I1001 16:42:55.184145 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-szkjs_2f53e292-fdef-4716-b217-0c4fab2dd64c/kube-rbac-proxy/0.log" Oct 01 16:42:55 crc kubenswrapper[4688]: I1001 16:42:55.215203 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-szkjs_2f53e292-fdef-4716-b217-0c4fab2dd64c/machine-api-operator/0.log" Oct 01 16:43:08 crc kubenswrapper[4688]: I1001 16:43:08.185120 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ndwpx_6a722b12-5da2-4154-af09-c291e4530375/cert-manager-controller/0.log" Oct 01 16:43:08 crc kubenswrapper[4688]: I1001 16:43:08.281107 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-wb49t_4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f/cert-manager-cainjector/0.log" Oct 01 16:43:08 crc kubenswrapper[4688]: I1001 16:43:08.402399 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-zxblg_6654a118-5437-4fb0-9c11-1ac4cf9dba79/cert-manager-webhook/0.log" Oct 01 16:43:20 crc kubenswrapper[4688]: I1001 16:43:20.632670 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-dwkqk_1be1c4ac-558b-4b81-97cf-feb86083dd5b/nmstate-console-plugin/0.log" Oct 01 16:43:20 crc kubenswrapper[4688]: I1001 16:43:20.773783 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mjqzm_7e9edca6-ee2b-4366-be0a-7c974e4865d7/nmstate-handler/0.log" Oct 01 16:43:20 crc kubenswrapper[4688]: I1001 16:43:20.839196 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2x5vh_8e78e6a0-4428-40b7-b819-6d798cf9983b/kube-rbac-proxy/0.log" Oct 01 16:43:20 crc kubenswrapper[4688]: I1001 16:43:20.864744 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2x5vh_8e78e6a0-4428-40b7-b819-6d798cf9983b/nmstate-metrics/0.log" Oct 01 16:43:21 crc kubenswrapper[4688]: I1001 16:43:21.035791 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-mv2bb_bd754ddd-7d01-499a-9bfc-448c200a8c93/nmstate-operator/0.log" Oct 01 16:43:21 crc kubenswrapper[4688]: I1001 16:43:21.115905 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-x9vxk_67f45b7a-07fb-4cf3-b63d-832139f2bccd/nmstate-webhook/0.log" Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.052937 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.053469 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.053512 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.054412 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.054454 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a" gracePeriod=600 Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.531911 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a" exitCode=0 Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.531989 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a"} Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.532265 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137"} Oct 01 16:43:25 crc kubenswrapper[4688]: I1001 16:43:25.532290 4688 scope.go:117] "RemoveContainer" containerID="7856923abd4305e1819765096fdf6cb37d92036d1922a938fe9487fd7bd2b44b" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.724494 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:26 crc kubenswrapper[4688]: E1001 16:43:26.725283 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ece00f-3a01-48a5-9a3a-1f34b7834234" containerName="container-00" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.725301 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ece00f-3a01-48a5-9a3a-1f34b7834234" containerName="container-00" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.725605 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ece00f-3a01-48a5-9a3a-1f34b7834234" containerName="container-00" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.727281 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.734809 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.913127 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.913457 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:26 crc kubenswrapper[4688]: I1001 16:43:26.913509 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv9qf\" (UniqueName: \"kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.018480 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.018592 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv9qf\" (UniqueName: \"kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.018732 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.019233 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.019263 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.040622 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv9qf\" (UniqueName: \"kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf\") pod \"certified-operators-v4qtr\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.079629 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:27 crc kubenswrapper[4688]: I1001 16:43:27.786737 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:27 crc kubenswrapper[4688]: W1001 16:43:27.788355 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod311cd214_8560_4cf9_8c3a_6f32d0b0b51d.slice/crio-6cb0c455098f5bd0c30b5a7681c9a78e61e34acf540c5a5f12b50876b780d686 WatchSource:0}: Error finding container 6cb0c455098f5bd0c30b5a7681c9a78e61e34acf540c5a5f12b50876b780d686: Status 404 returned error can't find the container with id 6cb0c455098f5bd0c30b5a7681c9a78e61e34acf540c5a5f12b50876b780d686 Oct 01 16:43:28 crc kubenswrapper[4688]: I1001 16:43:28.562283 4688 generic.go:334] "Generic (PLEG): container finished" podID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerID="e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6" exitCode=0 Oct 01 16:43:28 crc kubenswrapper[4688]: I1001 16:43:28.562371 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerDied","Data":"e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6"} Oct 01 16:43:28 crc kubenswrapper[4688]: I1001 16:43:28.562562 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerStarted","Data":"6cb0c455098f5bd0c30b5a7681c9a78e61e34acf540c5a5f12b50876b780d686"} Oct 01 16:43:28 crc kubenswrapper[4688]: I1001 16:43:28.563998 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:43:30 crc kubenswrapper[4688]: I1001 16:43:30.585238 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerStarted","Data":"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d"} Oct 01 16:43:31 crc kubenswrapper[4688]: I1001 16:43:31.594836 4688 generic.go:334] "Generic (PLEG): container finished" podID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerID="80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d" exitCode=0 Oct 01 16:43:31 crc kubenswrapper[4688]: I1001 16:43:31.594910 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerDied","Data":"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d"} Oct 01 16:43:32 crc kubenswrapper[4688]: I1001 16:43:32.610370 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerStarted","Data":"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4"} Oct 01 16:43:32 crc kubenswrapper[4688]: I1001 16:43:32.657643 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4qtr" podStartSLOduration=3.202745239 podStartE2EDuration="6.657619975s" podCreationTimestamp="2025-10-01 16:43:26 +0000 UTC" firstStartedPulling="2025-10-01 16:43:28.563787215 +0000 UTC m=+3397.914427177" lastFinishedPulling="2025-10-01 16:43:32.018661961 +0000 UTC m=+3401.369301913" observedRunningTime="2025-10-01 16:43:32.649927296 +0000 UTC m=+3402.000567258" watchObservedRunningTime="2025-10-01 16:43:32.657619975 +0000 UTC m=+3402.008259957" Oct 01 16:43:35 crc kubenswrapper[4688]: I1001 16:43:35.801402 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-st9jl_57161b05-13a8-415e-b9e9-ab575ff86007/kube-rbac-proxy/0.log" Oct 01 16:43:35 crc kubenswrapper[4688]: I1001 16:43:35.894571 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-st9jl_57161b05-13a8-415e-b9e9-ab575ff86007/controller/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.015175 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.233811 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.242792 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.274946 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.296656 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.445301 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.484801 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.499085 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.523825 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.711616 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.733737 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.771536 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/controller/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.777454 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.932725 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/frr-metrics/0.log" Oct 01 16:43:36 crc kubenswrapper[4688]: I1001 16:43:36.963515 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/kube-rbac-proxy-frr/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.063761 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/kube-rbac-proxy/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.079714 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.079751 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.086471 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/frr/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.127614 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.193633 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/reloader/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.300586 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-s7mkk_b047ab98-f421-44b0-9ee5-d2affd7804b8/frr-k8s-webhook-server/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.392668 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6887bf55bf-g25xn_550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6/manager/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.498858 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-9b7c45d89-n5mmm_e8ccae4e-6cb9-40e8-b56b-4033f5d0f086/webhook-server/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.623138 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-j98l6_618070a7-23e1-4b85-a68e-85196f7838d9/kube-rbac-proxy/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.703292 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.715759 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-j98l6_618070a7-23e1-4b85-a68e-85196f7838d9/speaker/0.log" Oct 01 16:43:37 crc kubenswrapper[4688]: I1001 16:43:37.774471 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:39 crc kubenswrapper[4688]: I1001 16:43:39.661664 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4qtr" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="registry-server" containerID="cri-o://dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4" gracePeriod=2 Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.155693 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.257330 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content\") pod \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.257418 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv9qf\" (UniqueName: \"kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf\") pod \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.257614 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities\") pod \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\" (UID: \"311cd214-8560-4cf9-8c3a-6f32d0b0b51d\") " Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.258670 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities" (OuterVolumeSpecName: "utilities") pod "311cd214-8560-4cf9-8c3a-6f32d0b0b51d" (UID: "311cd214-8560-4cf9-8c3a-6f32d0b0b51d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.262387 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf" (OuterVolumeSpecName: "kube-api-access-tv9qf") pod "311cd214-8560-4cf9-8c3a-6f32d0b0b51d" (UID: "311cd214-8560-4cf9-8c3a-6f32d0b0b51d"). InnerVolumeSpecName "kube-api-access-tv9qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.307244 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "311cd214-8560-4cf9-8c3a-6f32d0b0b51d" (UID: "311cd214-8560-4cf9-8c3a-6f32d0b0b51d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.360314 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.360753 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv9qf\" (UniqueName: \"kubernetes.io/projected/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-kube-api-access-tv9qf\") on node \"crc\" DevicePath \"\"" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.360841 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311cd214-8560-4cf9-8c3a-6f32d0b0b51d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.673511 4688 generic.go:334] "Generic (PLEG): container finished" podID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerID="dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4" exitCode=0 Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.673601 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerDied","Data":"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4"} Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.673647 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4qtr" event={"ID":"311cd214-8560-4cf9-8c3a-6f32d0b0b51d","Type":"ContainerDied","Data":"6cb0c455098f5bd0c30b5a7681c9a78e61e34acf540c5a5f12b50876b780d686"} Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.673659 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4qtr" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.673683 4688 scope.go:117] "RemoveContainer" containerID="dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.703079 4688 scope.go:117] "RemoveContainer" containerID="80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.728022 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.728068 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4qtr"] Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.754379 4688 scope.go:117] "RemoveContainer" containerID="e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.793495 4688 scope.go:117] "RemoveContainer" containerID="dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4" Oct 01 16:43:40 crc kubenswrapper[4688]: E1001 16:43:40.793952 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4\": container with ID starting with dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4 not found: ID does not exist" containerID="dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.793994 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4"} err="failed to get container status \"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4\": rpc error: code = NotFound desc = could not find container \"dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4\": container with ID starting with dd8ebd359620caf68bb32eb581d49f4a27b0dff567932837f87bc4793b54afb4 not found: ID does not exist" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.794020 4688 scope.go:117] "RemoveContainer" containerID="80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d" Oct 01 16:43:40 crc kubenswrapper[4688]: E1001 16:43:40.794304 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d\": container with ID starting with 80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d not found: ID does not exist" containerID="80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.794333 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d"} err="failed to get container status \"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d\": rpc error: code = NotFound desc = could not find container \"80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d\": container with ID starting with 80aec6c044da4f4cbde0da6f743ac8f6f45620ef53ca5b7460e6ef890926195d not found: ID does not exist" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.794348 4688 scope.go:117] "RemoveContainer" containerID="e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6" Oct 01 16:43:40 crc kubenswrapper[4688]: E1001 16:43:40.794688 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6\": container with ID starting with e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6 not found: ID does not exist" containerID="e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6" Oct 01 16:43:40 crc kubenswrapper[4688]: I1001 16:43:40.794725 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6"} err="failed to get container status \"e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6\": rpc error: code = NotFound desc = could not find container \"e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6\": container with ID starting with e1e766333bdf9367586ac5cd0b3ee0cd91ed009b237e8bc70e0f0e01de4e3fc6 not found: ID does not exist" Oct 01 16:43:41 crc kubenswrapper[4688]: I1001 16:43:41.399278 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" path="/var/lib/kubelet/pods/311cd214-8560-4cf9-8c3a-6f32d0b0b51d/volumes" Oct 01 16:43:50 crc kubenswrapper[4688]: I1001 16:43:50.606703 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:43:50 crc kubenswrapper[4688]: I1001 16:43:50.776857 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:43:50 crc kubenswrapper[4688]: I1001 16:43:50.821249 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:43:50 crc kubenswrapper[4688]: I1001 16:43:50.841299 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.032599 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.040958 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.043473 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/extract/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.212436 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.364863 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.416831 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.437669 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.719205 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.778571 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/extract/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.793641 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:43:51 crc kubenswrapper[4688]: I1001 16:43:51.913277 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.110748 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.136194 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.137798 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.369886 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.420175 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.699340 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/registry-server/0.log" Oct 01 16:43:52 crc kubenswrapper[4688]: I1001 16:43:52.722016 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.088719 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.107620 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.208187 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.373317 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.518257 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:43:53 crc kubenswrapper[4688]: I1001 16:43:53.735383 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.019398 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/registry-server/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.038981 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.041498 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.042378 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.249872 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.316091 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/extract/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.342049 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.431566 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.667876 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.695035 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.748116 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.942113 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.957773 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kszs7_a368e405-b690-47e8-b2e7-ff9917147420/marketplace-operator/0.log" Oct 01 16:43:54 crc kubenswrapper[4688]: I1001 16:43:54.981545 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.010883 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/extract/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.181677 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.364675 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.378378 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.426663 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.538533 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.643173 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.706326 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/registry-server/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.739023 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.902388 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.934832 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:43:55 crc kubenswrapper[4688]: I1001 16:43:55.935955 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:43:56 crc kubenswrapper[4688]: I1001 16:43:56.093926 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:43:56 crc kubenswrapper[4688]: I1001 16:43:56.140600 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:43:56 crc kubenswrapper[4688]: I1001 16:43:56.325979 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/registry-server/0.log" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.223972 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd"] Oct 01 16:45:00 crc kubenswrapper[4688]: E1001 16:45:00.225234 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="extract-utilities" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.225261 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="extract-utilities" Oct 01 16:45:00 crc kubenswrapper[4688]: E1001 16:45:00.225283 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="registry-server" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.225294 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="registry-server" Oct 01 16:45:00 crc kubenswrapper[4688]: E1001 16:45:00.225340 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="extract-content" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.225354 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="extract-content" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.225692 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="311cd214-8560-4cf9-8c3a-6f32d0b0b51d" containerName="registry-server" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.226793 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.232062 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.232350 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.232965 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd"] Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.290259 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs5qd\" (UniqueName: \"kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.290388 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.290439 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.391967 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs5qd\" (UniqueName: \"kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.392105 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.392176 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.393580 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.410668 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.425648 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs5qd\" (UniqueName: \"kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd\") pod \"collect-profiles-29322285-m9jzd\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:00 crc kubenswrapper[4688]: I1001 16:45:00.569335 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:01 crc kubenswrapper[4688]: I1001 16:45:01.169084 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd"] Oct 01 16:45:01 crc kubenswrapper[4688]: I1001 16:45:01.509095 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" event={"ID":"52d90d41-21b4-4b44-aeac-56e7c0c71ec1","Type":"ContainerStarted","Data":"b213b449c4cb7c5f8bf81570b4e3d850448446be2f719e54571f8b29a6a18311"} Oct 01 16:45:01 crc kubenswrapper[4688]: I1001 16:45:01.509433 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" event={"ID":"52d90d41-21b4-4b44-aeac-56e7c0c71ec1","Type":"ContainerStarted","Data":"a9ac4747c0877c9d9bfce46b1cd191eec0267f18c98c22389e7277af8822ecd5"} Oct 01 16:45:01 crc kubenswrapper[4688]: I1001 16:45:01.525141 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" podStartSLOduration=1.525126555 podStartE2EDuration="1.525126555s" podCreationTimestamp="2025-10-01 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:45:01.524300771 +0000 UTC m=+3490.874940743" watchObservedRunningTime="2025-10-01 16:45:01.525126555 +0000 UTC m=+3490.875766527" Oct 01 16:45:02 crc kubenswrapper[4688]: I1001 16:45:02.522310 4688 generic.go:334] "Generic (PLEG): container finished" podID="52d90d41-21b4-4b44-aeac-56e7c0c71ec1" containerID="b213b449c4cb7c5f8bf81570b4e3d850448446be2f719e54571f8b29a6a18311" exitCode=0 Oct 01 16:45:02 crc kubenswrapper[4688]: I1001 16:45:02.523013 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" event={"ID":"52d90d41-21b4-4b44-aeac-56e7c0c71ec1","Type":"ContainerDied","Data":"b213b449c4cb7c5f8bf81570b4e3d850448446be2f719e54571f8b29a6a18311"} Oct 01 16:45:03 crc kubenswrapper[4688]: I1001 16:45:03.909152 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.103212 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume\") pod \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.103321 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs5qd\" (UniqueName: \"kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd\") pod \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.103419 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume\") pod \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\" (UID: \"52d90d41-21b4-4b44-aeac-56e7c0c71ec1\") " Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.104088 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume" (OuterVolumeSpecName: "config-volume") pod "52d90d41-21b4-4b44-aeac-56e7c0c71ec1" (UID: "52d90d41-21b4-4b44-aeac-56e7c0c71ec1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.109686 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd" (OuterVolumeSpecName: "kube-api-access-hs5qd") pod "52d90d41-21b4-4b44-aeac-56e7c0c71ec1" (UID: "52d90d41-21b4-4b44-aeac-56e7c0c71ec1"). InnerVolumeSpecName "kube-api-access-hs5qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.111161 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52d90d41-21b4-4b44-aeac-56e7c0c71ec1" (UID: "52d90d41-21b4-4b44-aeac-56e7c0c71ec1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.205635 4688 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.206008 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs5qd\" (UniqueName: \"kubernetes.io/projected/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-kube-api-access-hs5qd\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.206031 4688 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d90d41-21b4-4b44-aeac-56e7c0c71ec1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.496014 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc"] Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.504219 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322240-ssdlc"] Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.540060 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" event={"ID":"52d90d41-21b4-4b44-aeac-56e7c0c71ec1","Type":"ContainerDied","Data":"a9ac4747c0877c9d9bfce46b1cd191eec0267f18c98c22389e7277af8822ecd5"} Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.540107 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9ac4747c0877c9d9bfce46b1cd191eec0267f18c98c22389e7277af8822ecd5" Oct 01 16:45:04 crc kubenswrapper[4688]: I1001 16:45:04.540122 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322285-m9jzd" Oct 01 16:45:05 crc kubenswrapper[4688]: I1001 16:45:05.420221 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39498707-adfc-4c83-a1e3-da36568ff834" path="/var/lib/kubelet/pods/39498707-adfc-4c83-a1e3-da36568ff834/volumes" Oct 01 16:45:25 crc kubenswrapper[4688]: I1001 16:45:25.052990 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:45:25 crc kubenswrapper[4688]: I1001 16:45:25.053673 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:45:55 crc kubenswrapper[4688]: I1001 16:45:55.053423 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:45:55 crc kubenswrapper[4688]: I1001 16:45:55.055005 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:45:59 crc kubenswrapper[4688]: I1001 16:45:59.166591 4688 generic.go:334] "Generic (PLEG): container finished" podID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerID="2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35" exitCode=0 Oct 01 16:45:59 crc kubenswrapper[4688]: I1001 16:45:59.166655 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" event={"ID":"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2","Type":"ContainerDied","Data":"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35"} Oct 01 16:45:59 crc kubenswrapper[4688]: I1001 16:45:59.167959 4688 scope.go:117] "RemoveContainer" containerID="2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35" Oct 01 16:46:00 crc kubenswrapper[4688]: I1001 16:46:00.140144 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wzjvs_must-gather-9c2wh_a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2/gather/0.log" Oct 01 16:46:03 crc kubenswrapper[4688]: I1001 16:46:03.022951 4688 scope.go:117] "RemoveContainer" containerID="c6a5bbd0ec084639bcb0a1076a97c7d347ea8ebccfd66c85d592f77fba732b61" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.157556 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:07 crc kubenswrapper[4688]: E1001 16:46:07.158921 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d90d41-21b4-4b44-aeac-56e7c0c71ec1" containerName="collect-profiles" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.158939 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d90d41-21b4-4b44-aeac-56e7c0c71ec1" containerName="collect-profiles" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.159184 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d90d41-21b4-4b44-aeac-56e7c0c71ec1" containerName="collect-profiles" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.160879 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.172574 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.253240 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.253317 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.253362 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lng9m\" (UniqueName: \"kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.355020 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.355277 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.355400 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lng9m\" (UniqueName: \"kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.355668 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.355735 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.387007 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lng9m\" (UniqueName: \"kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m\") pod \"community-operators-zxgvg\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:07 crc kubenswrapper[4688]: I1001 16:46:07.532312 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.064450 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.295986 4688 generic.go:334] "Generic (PLEG): container finished" podID="499f7706-a0f8-467f-a456-13e848aaae3a" containerID="4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd" exitCode=0 Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.296023 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerDied","Data":"4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd"} Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.296051 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerStarted","Data":"d11d5b692bf9864287d83110f165ff05e4ea0b1005f0cd769822f36a8473df6d"} Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.380246 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wzjvs/must-gather-9c2wh"] Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.380507 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="copy" containerID="cri-o://56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6" gracePeriod=2 Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.388731 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wzjvs/must-gather-9c2wh"] Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.839369 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wzjvs_must-gather-9c2wh_a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2/copy/0.log" Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.840115 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.989709 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg5ts\" (UniqueName: \"kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts\") pod \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.989778 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output\") pod \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\" (UID: \"a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2\") " Oct 01 16:46:08 crc kubenswrapper[4688]: I1001 16:46:08.994507 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts" (OuterVolumeSpecName: "kube-api-access-zg5ts") pod "a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" (UID: "a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2"). InnerVolumeSpecName "kube-api-access-zg5ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.093326 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg5ts\" (UniqueName: \"kubernetes.io/projected/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-kube-api-access-zg5ts\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.156593 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" (UID: "a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.195939 4688 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.306671 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wzjvs_must-gather-9c2wh_a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2/copy/0.log" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.307371 4688 generic.go:334] "Generic (PLEG): container finished" podID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerID="56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6" exitCode=143 Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.307442 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wzjvs/must-gather-9c2wh" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.307455 4688 scope.go:117] "RemoveContainer" containerID="56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.336104 4688 scope.go:117] "RemoveContainer" containerID="2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.392700 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" path="/var/lib/kubelet/pods/a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2/volumes" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.397701 4688 scope.go:117] "RemoveContainer" containerID="56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6" Oct 01 16:46:09 crc kubenswrapper[4688]: E1001 16:46:09.398232 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6\": container with ID starting with 56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6 not found: ID does not exist" containerID="56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.398286 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6"} err="failed to get container status \"56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6\": rpc error: code = NotFound desc = could not find container \"56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6\": container with ID starting with 56e9739441ed7d58b9403f0355b9a302abc9c6b68122bda410c4fa3f1ffeb2f6 not found: ID does not exist" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.398311 4688 scope.go:117] "RemoveContainer" containerID="2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35" Oct 01 16:46:09 crc kubenswrapper[4688]: E1001 16:46:09.398627 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35\": container with ID starting with 2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35 not found: ID does not exist" containerID="2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35" Oct 01 16:46:09 crc kubenswrapper[4688]: I1001 16:46:09.398648 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35"} err="failed to get container status \"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35\": rpc error: code = NotFound desc = could not find container \"2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35\": container with ID starting with 2d9dfecfcabb8974fea292bc47a26abc819c74660efe1ca21182f87f0e467e35 not found: ID does not exist" Oct 01 16:46:10 crc kubenswrapper[4688]: I1001 16:46:10.319714 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerStarted","Data":"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc"} Oct 01 16:46:11 crc kubenswrapper[4688]: I1001 16:46:11.334423 4688 generic.go:334] "Generic (PLEG): container finished" podID="499f7706-a0f8-467f-a456-13e848aaae3a" containerID="3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc" exitCode=0 Oct 01 16:46:11 crc kubenswrapper[4688]: I1001 16:46:11.334471 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerDied","Data":"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc"} Oct 01 16:46:12 crc kubenswrapper[4688]: I1001 16:46:12.348634 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerStarted","Data":"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800"} Oct 01 16:46:12 crc kubenswrapper[4688]: I1001 16:46:12.380114 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zxgvg" podStartSLOduration=1.854015799 podStartE2EDuration="5.38009197s" podCreationTimestamp="2025-10-01 16:46:07 +0000 UTC" firstStartedPulling="2025-10-01 16:46:08.297585912 +0000 UTC m=+3557.648225874" lastFinishedPulling="2025-10-01 16:46:11.823662083 +0000 UTC m=+3561.174302045" observedRunningTime="2025-10-01 16:46:12.367851002 +0000 UTC m=+3561.718490964" watchObservedRunningTime="2025-10-01 16:46:12.38009197 +0000 UTC m=+3561.730731932" Oct 01 16:46:17 crc kubenswrapper[4688]: I1001 16:46:17.533089 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:17 crc kubenswrapper[4688]: I1001 16:46:17.533782 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:17 crc kubenswrapper[4688]: I1001 16:46:17.606720 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:18 crc kubenswrapper[4688]: I1001 16:46:18.454678 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:18 crc kubenswrapper[4688]: I1001 16:46:18.500143 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.425977 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zxgvg" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="registry-server" containerID="cri-o://cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800" gracePeriod=2 Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.827211 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.970426 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content\") pod \"499f7706-a0f8-467f-a456-13e848aaae3a\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.970542 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities\") pod \"499f7706-a0f8-467f-a456-13e848aaae3a\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.970609 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lng9m\" (UniqueName: \"kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m\") pod \"499f7706-a0f8-467f-a456-13e848aaae3a\" (UID: \"499f7706-a0f8-467f-a456-13e848aaae3a\") " Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.971393 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities" (OuterVolumeSpecName: "utilities") pod "499f7706-a0f8-467f-a456-13e848aaae3a" (UID: "499f7706-a0f8-467f-a456-13e848aaae3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:20 crc kubenswrapper[4688]: I1001 16:46:20.976837 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m" (OuterVolumeSpecName: "kube-api-access-lng9m") pod "499f7706-a0f8-467f-a456-13e848aaae3a" (UID: "499f7706-a0f8-467f-a456-13e848aaae3a"). InnerVolumeSpecName "kube-api-access-lng9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.041077 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "499f7706-a0f8-467f-a456-13e848aaae3a" (UID: "499f7706-a0f8-467f-a456-13e848aaae3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.072925 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.072958 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/499f7706-a0f8-467f-a456-13e848aaae3a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.072968 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lng9m\" (UniqueName: \"kubernetes.io/projected/499f7706-a0f8-467f-a456-13e848aaae3a-kube-api-access-lng9m\") on node \"crc\" DevicePath \"\"" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.437205 4688 generic.go:334] "Generic (PLEG): container finished" podID="499f7706-a0f8-467f-a456-13e848aaae3a" containerID="cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800" exitCode=0 Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.437273 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerDied","Data":"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800"} Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.437328 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxgvg" event={"ID":"499f7706-a0f8-467f-a456-13e848aaae3a","Type":"ContainerDied","Data":"d11d5b692bf9864287d83110f165ff05e4ea0b1005f0cd769822f36a8473df6d"} Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.437361 4688 scope.go:117] "RemoveContainer" containerID="cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.437617 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxgvg" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.463725 4688 scope.go:117] "RemoveContainer" containerID="3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.464471 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.474983 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zxgvg"] Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.490230 4688 scope.go:117] "RemoveContainer" containerID="4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.552802 4688 scope.go:117] "RemoveContainer" containerID="cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800" Oct 01 16:46:21 crc kubenswrapper[4688]: E1001 16:46:21.554061 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800\": container with ID starting with cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800 not found: ID does not exist" containerID="cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.554113 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800"} err="failed to get container status \"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800\": rpc error: code = NotFound desc = could not find container \"cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800\": container with ID starting with cc761e18d38daf7d751854d32fa870951633096c31ed997f043d2044141aa800 not found: ID does not exist" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.554139 4688 scope.go:117] "RemoveContainer" containerID="3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc" Oct 01 16:46:21 crc kubenswrapper[4688]: E1001 16:46:21.554693 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc\": container with ID starting with 3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc not found: ID does not exist" containerID="3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.554732 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc"} err="failed to get container status \"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc\": rpc error: code = NotFound desc = could not find container \"3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc\": container with ID starting with 3600d27f81ef9eb24003cba009351e9e0fe9dd9bf24b2968d12d717e24c87ffc not found: ID does not exist" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.554752 4688 scope.go:117] "RemoveContainer" containerID="4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd" Oct 01 16:46:21 crc kubenswrapper[4688]: E1001 16:46:21.563643 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd\": container with ID starting with 4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd not found: ID does not exist" containerID="4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd" Oct 01 16:46:21 crc kubenswrapper[4688]: I1001 16:46:21.563684 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd"} err="failed to get container status \"4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd\": rpc error: code = NotFound desc = could not find container \"4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd\": container with ID starting with 4ccf9c5317b3f3fbcb90333806f56a6df9531998fd5c4d120741b572057e31bd not found: ID does not exist" Oct 01 16:46:23 crc kubenswrapper[4688]: I1001 16:46:23.392601 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" path="/var/lib/kubelet/pods/499f7706-a0f8-467f-a456-13e848aaae3a/volumes" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.053346 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.054018 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.054194 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.056648 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.056886 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" gracePeriod=600 Oct 01 16:46:25 crc kubenswrapper[4688]: E1001 16:46:25.184999 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.489997 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" exitCode=0 Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.490076 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137"} Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.490137 4688 scope.go:117] "RemoveContainer" containerID="afb7811593725ca5e01ce763fa5a89e102e681c1fa5eab24fda17166c778904a" Oct 01 16:46:25 crc kubenswrapper[4688]: I1001 16:46:25.491088 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:46:25 crc kubenswrapper[4688]: E1001 16:46:25.491634 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:46:37 crc kubenswrapper[4688]: I1001 16:46:37.381475 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:46:37 crc kubenswrapper[4688]: E1001 16:46:37.382790 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.754785 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fbcdq/must-gather-xkt8s"] Oct 01 16:46:47 crc kubenswrapper[4688]: E1001 16:46:47.755639 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="extract-utilities" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755652 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="extract-utilities" Oct 01 16:46:47 crc kubenswrapper[4688]: E1001 16:46:47.755663 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="copy" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755669 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="copy" Oct 01 16:46:47 crc kubenswrapper[4688]: E1001 16:46:47.755687 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="extract-content" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755694 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="extract-content" Oct 01 16:46:47 crc kubenswrapper[4688]: E1001 16:46:47.755714 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="registry-server" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755720 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="registry-server" Oct 01 16:46:47 crc kubenswrapper[4688]: E1001 16:46:47.755739 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="gather" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755745 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="gather" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755939 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="gather" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755958 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4954410-a9e2-4f72-a18e-ce2ef3cdf2b2" containerName="copy" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.755975 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="499f7706-a0f8-467f-a456-13e848aaae3a" containerName="registry-server" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.757044 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.759833 4688 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-fbcdq"/"default-dockercfg-2xjwt" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.759832 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fbcdq"/"openshift-service-ca.crt" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.760162 4688 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fbcdq"/"kube-root-ca.crt" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.765948 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fbcdq/must-gather-xkt8s"] Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.843752 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fh4t\" (UniqueName: \"kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.844088 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.946086 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.946230 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fh4t\" (UniqueName: \"kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.946713 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:47 crc kubenswrapper[4688]: I1001 16:46:47.976716 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fh4t\" (UniqueName: \"kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t\") pod \"must-gather-xkt8s\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:48 crc kubenswrapper[4688]: I1001 16:46:48.074050 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:46:48 crc kubenswrapper[4688]: I1001 16:46:48.531734 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fbcdq/must-gather-xkt8s"] Oct 01 16:46:48 crc kubenswrapper[4688]: I1001 16:46:48.730435 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" event={"ID":"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0","Type":"ContainerStarted","Data":"198a4da69446b6a0954c067a0e535fa19ba2f186be2bbf7b540a808ac6b0ef5a"} Oct 01 16:46:49 crc kubenswrapper[4688]: I1001 16:46:49.770145 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" event={"ID":"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0","Type":"ContainerStarted","Data":"c8fe70c70373ac8500bf84b0e92bf1f6cd02b6d6ed38353f5e857bdf9aaa18fe"} Oct 01 16:46:49 crc kubenswrapper[4688]: I1001 16:46:49.770388 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" event={"ID":"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0","Type":"ContainerStarted","Data":"646de5c380721cd02718ac8abfd76c4ffe1cdae078131c3fe91f5ea06a4d77e4"} Oct 01 16:46:49 crc kubenswrapper[4688]: I1001 16:46:49.788199 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" podStartSLOduration=2.788182317 podStartE2EDuration="2.788182317s" podCreationTimestamp="2025-10-01 16:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:46:49.785885972 +0000 UTC m=+3599.136525924" watchObservedRunningTime="2025-10-01 16:46:49.788182317 +0000 UTC m=+3599.138822279" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.381322 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:46:52 crc kubenswrapper[4688]: E1001 16:46:52.382705 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.540839 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-2cmvl"] Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.543220 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.638951 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.639034 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpdz8\" (UniqueName: \"kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.741455 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.741556 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpdz8\" (UniqueName: \"kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.742032 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.761716 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpdz8\" (UniqueName: \"kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8\") pod \"crc-debug-2cmvl\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: I1001 16:46:52.858195 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:46:52 crc kubenswrapper[4688]: W1001 16:46:52.891384 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda344db6_8acb_4b42_bbb6_fef17d37f0dd.slice/crio-eed2e9a989ffd157c83a4b560722ac1790a23495da25f7069d7a80db141b7831 WatchSource:0}: Error finding container eed2e9a989ffd157c83a4b560722ac1790a23495da25f7069d7a80db141b7831: Status 404 returned error can't find the container with id eed2e9a989ffd157c83a4b560722ac1790a23495da25f7069d7a80db141b7831 Oct 01 16:46:53 crc kubenswrapper[4688]: I1001 16:46:53.810650 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" event={"ID":"da344db6-8acb-4b42-bbb6-fef17d37f0dd","Type":"ContainerStarted","Data":"04f959d4a9d42ca22a17d4f16c478707022ea230c2b77f8486417716f359f1c0"} Oct 01 16:46:53 crc kubenswrapper[4688]: I1001 16:46:53.812005 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" event={"ID":"da344db6-8acb-4b42-bbb6-fef17d37f0dd","Type":"ContainerStarted","Data":"eed2e9a989ffd157c83a4b560722ac1790a23495da25f7069d7a80db141b7831"} Oct 01 16:46:53 crc kubenswrapper[4688]: I1001 16:46:53.832811 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" podStartSLOduration=1.832774706 podStartE2EDuration="1.832774706s" podCreationTimestamp="2025-10-01 16:46:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:46:53.823683578 +0000 UTC m=+3603.174323550" watchObservedRunningTime="2025-10-01 16:46:53.832774706 +0000 UTC m=+3603.183414688" Oct 01 16:47:03 crc kubenswrapper[4688]: I1001 16:47:03.133561 4688 scope.go:117] "RemoveContainer" containerID="d9e033b62daa6b5716fa17780c48cc1a30714760b9f6016a9b5f2b2dee355eb2" Oct 01 16:47:03 crc kubenswrapper[4688]: I1001 16:47:03.380920 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:47:03 crc kubenswrapper[4688]: E1001 16:47:03.381238 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:47:14 crc kubenswrapper[4688]: I1001 16:47:14.380510 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:47:14 crc kubenswrapper[4688]: E1001 16:47:14.381163 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:47:27 crc kubenswrapper[4688]: I1001 16:47:27.381618 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:47:27 crc kubenswrapper[4688]: E1001 16:47:27.382325 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:47:39 crc kubenswrapper[4688]: I1001 16:47:39.381825 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:47:39 crc kubenswrapper[4688]: E1001 16:47:39.382388 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:47:51 crc kubenswrapper[4688]: I1001 16:47:51.390954 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:47:51 crc kubenswrapper[4688]: E1001 16:47:51.398785 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.345733 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8668d66c66-j8wfd_6de36fb1-5b60-4600-ae35-bd31fdec91ce/barbican-api/0.log" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.381368 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:48:03 crc kubenswrapper[4688]: E1001 16:48:03.381575 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.455044 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8668d66c66-j8wfd_6de36fb1-5b60-4600-ae35-bd31fdec91ce/barbican-api-log/0.log" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.577313 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-df57dbd74-8gcrt_d13b3026-f303-454c-a686-7bc22f5211cf/barbican-keystone-listener/0.log" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.849383 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-df57dbd74-8gcrt_d13b3026-f303-454c-a686-7bc22f5211cf/barbican-keystone-listener-log/0.log" Oct 01 16:48:03 crc kubenswrapper[4688]: I1001 16:48:03.945351 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74574df4f9-v4dnz_b07391ac-b11b-4b46-8e9b-847b424e7a80/barbican-worker/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.068402 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74574df4f9-v4dnz_b07391ac-b11b-4b46-8e9b-847b424e7a80/barbican-worker-log/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.183913 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bk825_fbd7ce66-5861-4a86-a892-ce6493baf848/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.349252 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/ceilometer-central-agent/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.490148 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/ceilometer-notification-agent/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.598901 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/sg-core/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.640256 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ab269f92-11bd-4ea1-82a3-a77c15c5bb07/proxy-httpd/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.837001 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_12a98d94-2c07-4fa8-b69b-c24a11fc6cac/cinder-api-log/0.log" Oct 01 16:48:04 crc kubenswrapper[4688]: I1001 16:48:04.910380 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_12a98d94-2c07-4fa8-b69b-c24a11fc6cac/cinder-api/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.145499 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_825e667e-354c-4cd3-bb07-d51c67981ac8/cinder-scheduler/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.151902 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_825e667e-354c-4cd3-bb07-d51c67981ac8/probe/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.359853 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k4lch_3fb9e457-e82c-4418-a5a6-a3892a359683/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.594617 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7h7v6_42f771c8-2902-4358-8f27-3dd4e9ac510e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.724640 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-t2cfk_4633912c-f539-44fd-8fc4-1e57ddbdad56/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:05 crc kubenswrapper[4688]: I1001 16:48:05.953107 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/init/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.166444 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/init/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.268421 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b6dc74c5-vhf77_3b618166-5602-4785-8893-eb7b86fbe39a/dnsmasq-dns/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.426983 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-k9k2f_9cd2ef48-3ade-4edb-bd61-c3449a9c64fe/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.601448 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_32ec1a41-4380-4f29-a56d-dd2d722999d6/glance-httpd/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.631190 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_32ec1a41-4380-4f29-a56d-dd2d722999d6/glance-log/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.797762 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ab7d6e5f-ceab-400c-858e-2059dc492923/glance-httpd/0.log" Oct 01 16:48:06 crc kubenswrapper[4688]: I1001 16:48:06.889504 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ab7d6e5f-ceab-400c-858e-2059dc492923/glance-log/0.log" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.153340 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.155354 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.190050 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjczk\" (UniqueName: \"kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.190343 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.190442 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.214418 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.269542 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon/1.log" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.291962 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjczk\" (UniqueName: \"kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.292203 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.292311 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.293023 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.293360 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.327320 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjczk\" (UniqueName: \"kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk\") pod \"redhat-marketplace-7xjsb\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.440477 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon/0.log" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.579465 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.750007 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.755632 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.800687 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.912255 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.912556 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6clm\" (UniqueName: \"kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:07 crc kubenswrapper[4688]: I1001 16:48:07.912622 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.014974 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.015211 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.015359 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6clm\" (UniqueName: \"kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.016155 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.016430 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.027065 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9h8x_944f4da8-c2bb-400d-a78f-caf11336d5ec/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.050414 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6clm\" (UniqueName: \"kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm\") pod \"redhat-operators-lc6w7\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.096281 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-555656bf5b-xlntq_37923e92-dbcc-41a1-8d2f-89d8de59959e/horizon-log/0.log" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.110161 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.133485 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pdvcs_77d2978e-28f0-4591-a3fc-491ebfb6d1d9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.208719 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:08 crc kubenswrapper[4688]: W1001 16:48:08.223456 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd66207c7_946e_4e02_9b0a_5e31f641da93.slice/crio-b5170589ffb3222abf3e8d366627bfd6a093fe21a2770ba7b9c64bce9f9e6304 WatchSource:0}: Error finding container b5170589ffb3222abf3e8d366627bfd6a093fe21a2770ba7b9c64bce9f9e6304: Status 404 returned error can't find the container with id b5170589ffb3222abf3e8d366627bfd6a093fe21a2770ba7b9c64bce9f9e6304 Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.496251 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.496319 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c522fa53-fdbd-4836-b127-a858c5001cc0/kube-state-metrics/0.log" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.500811 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerStarted","Data":"b5170589ffb3222abf3e8d366627bfd6a093fe21a2770ba7b9c64bce9f9e6304"} Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.943666 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-587df57d87-hlpwt_8a94d646-b2be-4ab1-89ff-da8aa30591f6/keystone-api/0.log" Oct 01 16:48:08 crc kubenswrapper[4688]: I1001 16:48:08.995297 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hfj6x_959f38f8-3df7-422b-9c09-a6fa68d4b8f0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.322400 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c946d4bf9-l8zds_5766bcc5-cf95-400d-a634-c2e0ad4fe3b9/neutron-api/0.log" Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.371887 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c946d4bf9-l8zds_5766bcc5-cf95-400d-a634-c2e0ad4fe3b9/neutron-httpd/0.log" Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.535888 4688 generic.go:334] "Generic (PLEG): container finished" podID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerID="572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181" exitCode=0 Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.536050 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerDied","Data":"572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181"} Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.538977 4688 generic.go:334] "Generic (PLEG): container finished" podID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerID="05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760" exitCode=0 Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.539002 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerDied","Data":"05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760"} Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.539021 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerStarted","Data":"1610dd668ada6072f8d4ea6cf58787a0cf307b0b0f3d3dd169399884bec167e8"} Oct 01 16:48:09 crc kubenswrapper[4688]: I1001 16:48:09.695255 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wg5hz_0ddaaf82-b5e6-4960-b79e-77ad2f913b93/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.072743 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cdb337c8-732c-4123-b47b-212ecb38257d/nova-api-log/0.log" Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.444375 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_cdb337c8-732c-4123-b47b-212ecb38257d/nova-api-api/0.log" Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.502101 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f64f51ee-f2ae-4c33-babf-67838b0f0135/nova-cell0-conductor-conductor/0.log" Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.554307 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerStarted","Data":"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3"} Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.820562 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_72549a6c-a0a7-4c63-a8f5-efe5f6b6eac2/nova-cell1-conductor-conductor/0.log" Oct 01 16:48:10 crc kubenswrapper[4688]: I1001 16:48:10.953751 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_050f05f1-0f70-46e1-94a3-e32d2aa42c36/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 16:48:11 crc kubenswrapper[4688]: I1001 16:48:11.525820 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8g64m_6206a3c7-1bb2-422a-8bc9-38e62d890513/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:11 crc kubenswrapper[4688]: I1001 16:48:11.574215 4688 generic.go:334] "Generic (PLEG): container finished" podID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerID="ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3" exitCode=0 Oct 01 16:48:11 crc kubenswrapper[4688]: I1001 16:48:11.574591 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerDied","Data":"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3"} Oct 01 16:48:11 crc kubenswrapper[4688]: I1001 16:48:11.584626 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerStarted","Data":"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6"} Oct 01 16:48:11 crc kubenswrapper[4688]: I1001 16:48:11.635448 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c9f50e4d-f039-4902-ab15-03604f701daf/nova-metadata-log/0.log" Oct 01 16:48:12 crc kubenswrapper[4688]: I1001 16:48:12.173354 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8d577d2f-a1a6-4d70-8532-86c0bb41f51f/nova-scheduler-scheduler/0.log" Oct 01 16:48:12 crc kubenswrapper[4688]: I1001 16:48:12.440942 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/mysql-bootstrap/0.log" Oct 01 16:48:12 crc kubenswrapper[4688]: I1001 16:48:12.574404 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/mysql-bootstrap/0.log" Oct 01 16:48:12 crc kubenswrapper[4688]: I1001 16:48:12.788614 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_81865451-e9a4-4f89-aa96-eb345800ffe3/galera/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.099672 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/mysql-bootstrap/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.238014 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/mysql-bootstrap/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.282109 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c9f50e4d-f039-4902-ab15-03604f701daf/nova-metadata-metadata/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.448621 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_65ea9633-b867-4c49-8fb3-471793344f3b/galera/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.619478 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerStarted","Data":"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8"} Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.650643 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_a309f66f-cc9d-40e7-9acc-00881395f4c9/openstackclient/0.log" Oct 01 16:48:13 crc kubenswrapper[4688]: I1001 16:48:13.659780 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7xjsb" podStartSLOduration=3.432440514 podStartE2EDuration="6.659764079s" podCreationTimestamp="2025-10-01 16:48:07 +0000 UTC" firstStartedPulling="2025-10-01 16:48:09.538422427 +0000 UTC m=+3678.889062389" lastFinishedPulling="2025-10-01 16:48:12.765745992 +0000 UTC m=+3682.116385954" observedRunningTime="2025-10-01 16:48:13.645969227 +0000 UTC m=+3682.996609189" watchObservedRunningTime="2025-10-01 16:48:13.659764079 +0000 UTC m=+3683.010404041" Oct 01 16:48:14 crc kubenswrapper[4688]: I1001 16:48:14.091490 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sfjc5_9b3cd14b-091c-461a-8ce5-924f9619a596/openstack-network-exporter/0.log" Oct 01 16:48:14 crc kubenswrapper[4688]: I1001 16:48:14.179423 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server-init/0.log" Oct 01 16:48:14 crc kubenswrapper[4688]: I1001 16:48:14.639206 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server/0.log" Oct 01 16:48:14 crc kubenswrapper[4688]: I1001 16:48:14.647636 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovsdb-server-init/0.log" Oct 01 16:48:14 crc kubenswrapper[4688]: I1001 16:48:14.693503 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9wpf7_3ab65b16-215b-46d3-9fd7-98982427fffa/ovs-vswitchd/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.011807 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ts722_e4559e8a-674a-427d-a9fb-145f94ff3b58/ovn-controller/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.264445 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-t4nl2_da404969-614a-4c71-9ce3-0a15a702aa65/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.297094 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_33e5a114-5ed4-4df4-88ca-908340de2204/openstack-network-exporter/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.384923 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:48:15 crc kubenswrapper[4688]: E1001 16:48:15.385161 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.538015 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_33e5a114-5ed4-4df4-88ca-908340de2204/ovn-northd/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.585952 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79a0889a-6692-43c9-928b-13e06ccd6ca1/openstack-network-exporter/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.640943 4688 generic.go:334] "Generic (PLEG): container finished" podID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerID="ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6" exitCode=0 Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.640983 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerDied","Data":"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6"} Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.799636 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79a0889a-6692-43c9-928b-13e06ccd6ca1/ovsdbserver-nb/0.log" Oct 01 16:48:15 crc kubenswrapper[4688]: I1001 16:48:15.916774 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_466a92b3-31f5-41fb-8810-76e95dd7a332/openstack-network-exporter/0.log" Oct 01 16:48:16 crc kubenswrapper[4688]: I1001 16:48:16.104195 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_466a92b3-31f5-41fb-8810-76e95dd7a332/ovsdbserver-sb/0.log" Oct 01 16:48:16 crc kubenswrapper[4688]: I1001 16:48:16.417461 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-76d4fdc768-gl782_d97f10ed-cfae-4409-b6fd-ae047f584ac3/placement-api/0.log" Oct 01 16:48:16 crc kubenswrapper[4688]: I1001 16:48:16.589717 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-76d4fdc768-gl782_d97f10ed-cfae-4409-b6fd-ae047f584ac3/placement-log/0.log" Oct 01 16:48:16 crc kubenswrapper[4688]: I1001 16:48:16.653247 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerStarted","Data":"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1"} Oct 01 16:48:16 crc kubenswrapper[4688]: I1001 16:48:16.824162 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/setup-container/0.log" Oct 01 16:48:17 crc kubenswrapper[4688]: I1001 16:48:17.128323 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/setup-container/0.log" Oct 01 16:48:17 crc kubenswrapper[4688]: I1001 16:48:17.266127 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e95ed7d7-8fb2-4397-bc8a-7ce7bd355fbb/rabbitmq/0.log" Oct 01 16:48:17 crc kubenswrapper[4688]: I1001 16:48:17.579924 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:17 crc kubenswrapper[4688]: I1001 16:48:17.581098 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:17 crc kubenswrapper[4688]: I1001 16:48:17.667203 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/setup-container/0.log" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.111345 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.111751 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.145865 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/rabbitmq/0.log" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.157644 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d1876945-781a-41af-a495-bee2af872b98/setup-container/0.log" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.626783 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-7xjsb" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="registry-server" probeResult="failure" output=< Oct 01 16:48:18 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:48:18 crc kubenswrapper[4688]: > Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.747033 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-c755p_9cebafe1-24f0-4751-b691-8fd0bc09359f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:18 crc kubenswrapper[4688]: I1001 16:48:18.811248 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-7tdxj_1d630ed1-0ec0-4b50-aef4-b8a75aa67270/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:19 crc kubenswrapper[4688]: I1001 16:48:19.203353 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vxjsc_d8469466-3ce0-4d29-8d96-6cce137617fa/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:19 crc kubenswrapper[4688]: I1001 16:48:19.203896 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc6w7" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" probeResult="failure" output=< Oct 01 16:48:19 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:48:19 crc kubenswrapper[4688]: > Oct 01 16:48:19 crc kubenswrapper[4688]: I1001 16:48:19.585508 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-r7vc9_a1720cc4-8ce5-4d4a-b0f0-47afac314627/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:19 crc kubenswrapper[4688]: I1001 16:48:19.762947 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-bj9n4_289bca66-8942-4580-b92e-d5cb8db5deed/ssh-known-hosts-edpm-deployment/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.161288 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bbc5fdc47-8258p_43929b9b-990a-4a85-9100-509d80bba3c1/proxy-server/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.174971 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bbc5fdc47-8258p_43929b9b-990a-4a85-9100-509d80bba3c1/proxy-httpd/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.443241 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-96jrl_f47bc7a7-66e7-4994-985d-2ff7da317e55/swift-ring-rebalance/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.784962 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-auditor/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.821745 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-reaper/0.log" Oct 01 16:48:20 crc kubenswrapper[4688]: I1001 16:48:20.961495 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-replicator/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.098550 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-auditor/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.101275 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/account-server/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.294342 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-replicator/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.379178 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-updater/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.407381 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/container-server/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.641005 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-auditor/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.669595 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-replicator/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.696153 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-expirer/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.918409 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-server/0.log" Oct 01 16:48:21 crc kubenswrapper[4688]: I1001 16:48:21.998819 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/object-updater/0.log" Oct 01 16:48:22 crc kubenswrapper[4688]: I1001 16:48:22.072985 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/rsync/0.log" Oct 01 16:48:22 crc kubenswrapper[4688]: I1001 16:48:22.203769 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0cd57226-8163-4319-ad42-5a17df19881a/swift-recon-cron/0.log" Oct 01 16:48:22 crc kubenswrapper[4688]: I1001 16:48:22.502637 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wbrp2_149aba72-0ea1-4a18-9ff1-98cfd0ada453/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:22 crc kubenswrapper[4688]: I1001 16:48:22.646513 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_52eda021-1010-4c20-8a69-9b7df698f765/tempest-tests-tempest-tests-runner/0.log" Oct 01 16:48:22 crc kubenswrapper[4688]: I1001 16:48:22.865162 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_76b03e8a-66be-4233-99bd-3109365d180a/test-operator-logs-container/0.log" Oct 01 16:48:23 crc kubenswrapper[4688]: I1001 16:48:23.171539 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-whw8p_fd4cc96c-61fa-4f47-8a36-8d2c6ddef862/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 16:48:26 crc kubenswrapper[4688]: I1001 16:48:26.380890 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:48:26 crc kubenswrapper[4688]: E1001 16:48:26.381402 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:27 crc kubenswrapper[4688]: I1001 16:48:27.630221 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:27 crc kubenswrapper[4688]: I1001 16:48:27.665269 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lc6w7" podStartSLOduration=13.908044769 podStartE2EDuration="20.665251693s" podCreationTimestamp="2025-10-01 16:48:07 +0000 UTC" firstStartedPulling="2025-10-01 16:48:09.542068131 +0000 UTC m=+3678.892708083" lastFinishedPulling="2025-10-01 16:48:16.299275045 +0000 UTC m=+3685.649915007" observedRunningTime="2025-10-01 16:48:16.67885831 +0000 UTC m=+3686.029498272" watchObservedRunningTime="2025-10-01 16:48:27.665251693 +0000 UTC m=+3697.015891655" Oct 01 16:48:27 crc kubenswrapper[4688]: I1001 16:48:27.692320 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:27 crc kubenswrapper[4688]: I1001 16:48:27.880844 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:28 crc kubenswrapper[4688]: I1001 16:48:28.491762 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3720d335-ad5b-41cf-9acb-7f9b49fb2c6c/memcached/0.log" Oct 01 16:48:28 crc kubenswrapper[4688]: I1001 16:48:28.783021 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7xjsb" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="registry-server" containerID="cri-o://0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8" gracePeriod=2 Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.204075 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc6w7" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" probeResult="failure" output=< Oct 01 16:48:29 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:48:29 crc kubenswrapper[4688]: > Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.388101 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.497225 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjczk\" (UniqueName: \"kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk\") pod \"d66207c7-946e-4e02-9b0a-5e31f641da93\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.497498 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities\") pod \"d66207c7-946e-4e02-9b0a-5e31f641da93\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.497590 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content\") pod \"d66207c7-946e-4e02-9b0a-5e31f641da93\" (UID: \"d66207c7-946e-4e02-9b0a-5e31f641da93\") " Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.497902 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities" (OuterVolumeSpecName: "utilities") pod "d66207c7-946e-4e02-9b0a-5e31f641da93" (UID: "d66207c7-946e-4e02-9b0a-5e31f641da93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.498115 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.510699 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk" (OuterVolumeSpecName: "kube-api-access-kjczk") pod "d66207c7-946e-4e02-9b0a-5e31f641da93" (UID: "d66207c7-946e-4e02-9b0a-5e31f641da93"). InnerVolumeSpecName "kube-api-access-kjczk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.525737 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d66207c7-946e-4e02-9b0a-5e31f641da93" (UID: "d66207c7-946e-4e02-9b0a-5e31f641da93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.600151 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66207c7-946e-4e02-9b0a-5e31f641da93-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.600194 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjczk\" (UniqueName: \"kubernetes.io/projected/d66207c7-946e-4e02-9b0a-5e31f641da93-kube-api-access-kjczk\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.794023 4688 generic.go:334] "Generic (PLEG): container finished" podID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerID="0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8" exitCode=0 Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.794061 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerDied","Data":"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8"} Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.794085 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7xjsb" event={"ID":"d66207c7-946e-4e02-9b0a-5e31f641da93","Type":"ContainerDied","Data":"b5170589ffb3222abf3e8d366627bfd6a093fe21a2770ba7b9c64bce9f9e6304"} Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.794102 4688 scope.go:117] "RemoveContainer" containerID="0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.794230 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7xjsb" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.836542 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.845338 4688 scope.go:117] "RemoveContainer" containerID="ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.849043 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7xjsb"] Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.877216 4688 scope.go:117] "RemoveContainer" containerID="572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.916594 4688 scope.go:117] "RemoveContainer" containerID="0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8" Oct 01 16:48:29 crc kubenswrapper[4688]: E1001 16:48:29.917063 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8\": container with ID starting with 0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8 not found: ID does not exist" containerID="0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.917116 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8"} err="failed to get container status \"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8\": rpc error: code = NotFound desc = could not find container \"0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8\": container with ID starting with 0ba47a9c8063bad1721d1883dfe4019c39ad660534dac5970a95b1df916bd0e8 not found: ID does not exist" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.917149 4688 scope.go:117] "RemoveContainer" containerID="ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3" Oct 01 16:48:29 crc kubenswrapper[4688]: E1001 16:48:29.917695 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3\": container with ID starting with ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3 not found: ID does not exist" containerID="ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.917741 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3"} err="failed to get container status \"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3\": rpc error: code = NotFound desc = could not find container \"ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3\": container with ID starting with ec028fbf1b53d64135c9da533ef0010d122508178395dbcd91bbd1107f44f4c3 not found: ID does not exist" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.917769 4688 scope.go:117] "RemoveContainer" containerID="572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181" Oct 01 16:48:29 crc kubenswrapper[4688]: E1001 16:48:29.919825 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181\": container with ID starting with 572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181 not found: ID does not exist" containerID="572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181" Oct 01 16:48:29 crc kubenswrapper[4688]: I1001 16:48:29.919860 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181"} err="failed to get container status \"572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181\": rpc error: code = NotFound desc = could not find container \"572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181\": container with ID starting with 572e3631cb178b9686ca8a51ed7e114a8bcae26374cd57883cd0647dc0385181 not found: ID does not exist" Oct 01 16:48:31 crc kubenswrapper[4688]: I1001 16:48:31.390304 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" path="/var/lib/kubelet/pods/d66207c7-946e-4e02-9b0a-5e31f641da93/volumes" Oct 01 16:48:39 crc kubenswrapper[4688]: I1001 16:48:39.154326 4688 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc6w7" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" probeResult="failure" output=< Oct 01 16:48:39 crc kubenswrapper[4688]: timeout: failed to connect service ":50051" within 1s Oct 01 16:48:39 crc kubenswrapper[4688]: > Oct 01 16:48:41 crc kubenswrapper[4688]: I1001 16:48:41.392578 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:48:41 crc kubenswrapper[4688]: E1001 16:48:41.393198 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:48 crc kubenswrapper[4688]: I1001 16:48:48.159308 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:48 crc kubenswrapper[4688]: I1001 16:48:48.220499 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:49 crc kubenswrapper[4688]: I1001 16:48:49.398995 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:49 crc kubenswrapper[4688]: I1001 16:48:49.973065 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lc6w7" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" containerID="cri-o://2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1" gracePeriod=2 Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.436177 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.605672 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6clm\" (UniqueName: \"kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm\") pod \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.605959 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities\") pod \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.606076 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content\") pod \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\" (UID: \"b2b8c21d-a5bb-4100-a32f-7c919e600d57\") " Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.606680 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities" (OuterVolumeSpecName: "utilities") pod "b2b8c21d-a5bb-4100-a32f-7c919e600d57" (UID: "b2b8c21d-a5bb-4100-a32f-7c919e600d57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.606884 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.620905 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm" (OuterVolumeSpecName: "kube-api-access-x6clm") pod "b2b8c21d-a5bb-4100-a32f-7c919e600d57" (UID: "b2b8c21d-a5bb-4100-a32f-7c919e600d57"). InnerVolumeSpecName "kube-api-access-x6clm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.686742 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2b8c21d-a5bb-4100-a32f-7c919e600d57" (UID: "b2b8c21d-a5bb-4100-a32f-7c919e600d57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.708964 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6clm\" (UniqueName: \"kubernetes.io/projected/b2b8c21d-a5bb-4100-a32f-7c919e600d57-kube-api-access-x6clm\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:50 crc kubenswrapper[4688]: I1001 16:48:50.708996 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b8c21d-a5bb-4100-a32f-7c919e600d57-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:50.998278 4688 generic.go:334] "Generic (PLEG): container finished" podID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerID="2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1" exitCode=0 Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:50.998331 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerDied","Data":"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1"} Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:50.998365 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc6w7" event={"ID":"b2b8c21d-a5bb-4100-a32f-7c919e600d57","Type":"ContainerDied","Data":"1610dd668ada6072f8d4ea6cf58787a0cf307b0b0f3d3dd169399884bec167e8"} Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:50.998392 4688 scope.go:117] "RemoveContainer" containerID="2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:50.998593 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc6w7" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.031970 4688 scope.go:117] "RemoveContainer" containerID="ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.064765 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.073269 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lc6w7"] Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.081868 4688 scope.go:117] "RemoveContainer" containerID="05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.124945 4688 scope.go:117] "RemoveContainer" containerID="2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1" Oct 01 16:48:51 crc kubenswrapper[4688]: E1001 16:48:51.128768 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1\": container with ID starting with 2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1 not found: ID does not exist" containerID="2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.128937 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1"} err="failed to get container status \"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1\": rpc error: code = NotFound desc = could not find container \"2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1\": container with ID starting with 2a1ceb975100a4bdd543e0b0c414d8f3493bf0ca709badfe21759644b3a001e1 not found: ID does not exist" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.129145 4688 scope.go:117] "RemoveContainer" containerID="ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6" Oct 01 16:48:51 crc kubenswrapper[4688]: E1001 16:48:51.129572 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6\": container with ID starting with ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6 not found: ID does not exist" containerID="ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.129594 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6"} err="failed to get container status \"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6\": rpc error: code = NotFound desc = could not find container \"ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6\": container with ID starting with ba6d9c6c2c774ae61c34903ac7885dd5243a9dea95730dd98d5146021e49a4b6 not found: ID does not exist" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.129608 4688 scope.go:117] "RemoveContainer" containerID="05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760" Oct 01 16:48:51 crc kubenswrapper[4688]: E1001 16:48:51.129834 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760\": container with ID starting with 05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760 not found: ID does not exist" containerID="05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.129849 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760"} err="failed to get container status \"05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760\": rpc error: code = NotFound desc = could not find container \"05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760\": container with ID starting with 05ceb721250cb02860e81965ca86d7e2e4b90a2d74e18532c18b17ab53e14760 not found: ID does not exist" Oct 01 16:48:51 crc kubenswrapper[4688]: I1001 16:48:51.427091 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" path="/var/lib/kubelet/pods/b2b8c21d-a5bb-4100-a32f-7c919e600d57/volumes" Oct 01 16:48:52 crc kubenswrapper[4688]: I1001 16:48:52.380625 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:48:52 crc kubenswrapper[4688]: E1001 16:48:52.381142 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:48:53 crc kubenswrapper[4688]: I1001 16:48:53.047817 4688 generic.go:334] "Generic (PLEG): container finished" podID="da344db6-8acb-4b42-bbb6-fef17d37f0dd" containerID="04f959d4a9d42ca22a17d4f16c478707022ea230c2b77f8486417716f359f1c0" exitCode=0 Oct 01 16:48:53 crc kubenswrapper[4688]: I1001 16:48:53.047864 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" event={"ID":"da344db6-8acb-4b42-bbb6-fef17d37f0dd","Type":"ContainerDied","Data":"04f959d4a9d42ca22a17d4f16c478707022ea230c2b77f8486417716f359f1c0"} Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.161697 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.188318 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host\") pod \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.188458 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host" (OuterVolumeSpecName: "host") pod "da344db6-8acb-4b42-bbb6-fef17d37f0dd" (UID: "da344db6-8acb-4b42-bbb6-fef17d37f0dd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.188476 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpdz8\" (UniqueName: \"kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8\") pod \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\" (UID: \"da344db6-8acb-4b42-bbb6-fef17d37f0dd\") " Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.189319 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da344db6-8acb-4b42-bbb6-fef17d37f0dd-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.194892 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8" (OuterVolumeSpecName: "kube-api-access-kpdz8") pod "da344db6-8acb-4b42-bbb6-fef17d37f0dd" (UID: "da344db6-8acb-4b42-bbb6-fef17d37f0dd"). InnerVolumeSpecName "kube-api-access-kpdz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.202013 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-2cmvl"] Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.213604 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-2cmvl"] Oct 01 16:48:54 crc kubenswrapper[4688]: I1001 16:48:54.291139 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpdz8\" (UniqueName: \"kubernetes.io/projected/da344db6-8acb-4b42-bbb6-fef17d37f0dd-kube-api-access-kpdz8\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.070980 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eed2e9a989ffd157c83a4b560722ac1790a23495da25f7069d7a80db141b7831" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.071296 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-2cmvl" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.365908 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-xt86x"] Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366354 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da344db6-8acb-4b42-bbb6-fef17d37f0dd" containerName="container-00" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366372 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="da344db6-8acb-4b42-bbb6-fef17d37f0dd" containerName="container-00" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366411 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="extract-content" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366424 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="extract-content" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366439 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="extract-utilities" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366448 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="extract-utilities" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366468 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366476 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366506 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="extract-content" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366514 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="extract-content" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366557 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366566 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: E1001 16:48:55.366586 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="extract-utilities" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366595 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="extract-utilities" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366877 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b8c21d-a5bb-4100-a32f-7c919e600d57" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366909 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66207c7-946e-4e02-9b0a-5e31f641da93" containerName="registry-server" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.366923 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="da344db6-8acb-4b42-bbb6-fef17d37f0dd" containerName="container-00" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.367656 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.392246 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da344db6-8acb-4b42-bbb6-fef17d37f0dd" path="/var/lib/kubelet/pods/da344db6-8acb-4b42-bbb6-fef17d37f0dd/volumes" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.409687 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.409732 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xfll\" (UniqueName: \"kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.511372 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xfll\" (UniqueName: \"kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.511790 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.511899 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.533991 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xfll\" (UniqueName: \"kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll\") pod \"crc-debug-xt86x\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: I1001 16:48:55.688690 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:55 crc kubenswrapper[4688]: W1001 16:48:55.723278 4688 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2626c45_3a6f_44e7_aed2_74f1042b90f7.slice/crio-3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73 WatchSource:0}: Error finding container 3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73: Status 404 returned error can't find the container with id 3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73 Oct 01 16:48:56 crc kubenswrapper[4688]: I1001 16:48:56.084835 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" event={"ID":"a2626c45-3a6f-44e7-aed2-74f1042b90f7","Type":"ContainerStarted","Data":"32a0d78a9e94b1c5a7928ed23dfe36a1afaddf6ee6e73dc46d17c7d49eaf48e3"} Oct 01 16:48:56 crc kubenswrapper[4688]: I1001 16:48:56.085156 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" event={"ID":"a2626c45-3a6f-44e7-aed2-74f1042b90f7","Type":"ContainerStarted","Data":"3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73"} Oct 01 16:48:56 crc kubenswrapper[4688]: I1001 16:48:56.125498 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" podStartSLOduration=1.1254788279999999 podStartE2EDuration="1.125478828s" podCreationTimestamp="2025-10-01 16:48:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 16:48:56.099338394 +0000 UTC m=+3725.449978356" watchObservedRunningTime="2025-10-01 16:48:56.125478828 +0000 UTC m=+3725.476118790" Oct 01 16:48:57 crc kubenswrapper[4688]: I1001 16:48:57.100126 4688 generic.go:334] "Generic (PLEG): container finished" podID="a2626c45-3a6f-44e7-aed2-74f1042b90f7" containerID="32a0d78a9e94b1c5a7928ed23dfe36a1afaddf6ee6e73dc46d17c7d49eaf48e3" exitCode=0 Oct 01 16:48:57 crc kubenswrapper[4688]: I1001 16:48:57.101374 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" event={"ID":"a2626c45-3a6f-44e7-aed2-74f1042b90f7","Type":"ContainerDied","Data":"32a0d78a9e94b1c5a7928ed23dfe36a1afaddf6ee6e73dc46d17c7d49eaf48e3"} Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.216722 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.263383 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host\") pod \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.263483 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xfll\" (UniqueName: \"kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll\") pod \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\" (UID: \"a2626c45-3a6f-44e7-aed2-74f1042b90f7\") " Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.265234 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host" (OuterVolumeSpecName: "host") pod "a2626c45-3a6f-44e7-aed2-74f1042b90f7" (UID: "a2626c45-3a6f-44e7-aed2-74f1042b90f7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.271020 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll" (OuterVolumeSpecName: "kube-api-access-7xfll") pod "a2626c45-3a6f-44e7-aed2-74f1042b90f7" (UID: "a2626c45-3a6f-44e7-aed2-74f1042b90f7"). InnerVolumeSpecName "kube-api-access-7xfll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.365379 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2626c45-3a6f-44e7-aed2-74f1042b90f7-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:58 crc kubenswrapper[4688]: I1001 16:48:58.365406 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xfll\" (UniqueName: \"kubernetes.io/projected/a2626c45-3a6f-44e7-aed2-74f1042b90f7-kube-api-access-7xfll\") on node \"crc\" DevicePath \"\"" Oct 01 16:48:59 crc kubenswrapper[4688]: I1001 16:48:59.117879 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" event={"ID":"a2626c45-3a6f-44e7-aed2-74f1042b90f7","Type":"ContainerDied","Data":"3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73"} Oct 01 16:48:59 crc kubenswrapper[4688]: I1001 16:48:59.118191 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3200d82bb1b20d6d0c70b7e1728b9f8447c9d50b75fdc4c5a7009643c384ec73" Oct 01 16:48:59 crc kubenswrapper[4688]: I1001 16:48:59.118254 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-xt86x" Oct 01 16:49:02 crc kubenswrapper[4688]: I1001 16:49:02.064974 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-xt86x"] Oct 01 16:49:02 crc kubenswrapper[4688]: I1001 16:49:02.076196 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-xt86x"] Oct 01 16:49:03 crc kubenswrapper[4688]: I1001 16:49:03.291127 4688 scope.go:117] "RemoveContainer" containerID="51c65dbb5b42c526a743f2e9fa40c6ef3a3d3793fabaf8427a0d6ea93937232f" Oct 01 16:49:03 crc kubenswrapper[4688]: I1001 16:49:03.400779 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2626c45-3a6f-44e7-aed2-74f1042b90f7" path="/var/lib/kubelet/pods/a2626c45-3a6f-44e7-aed2-74f1042b90f7/volumes" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.359545 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-h4xzx"] Oct 01 16:49:04 crc kubenswrapper[4688]: E1001 16:49:04.359959 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2626c45-3a6f-44e7-aed2-74f1042b90f7" containerName="container-00" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.359976 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2626c45-3a6f-44e7-aed2-74f1042b90f7" containerName="container-00" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.360234 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2626c45-3a6f-44e7-aed2-74f1042b90f7" containerName="container-00" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.360952 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.499798 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zfsd\" (UniqueName: \"kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.500166 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.601698 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zfsd\" (UniqueName: \"kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.601747 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.601864 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.621106 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zfsd\" (UniqueName: \"kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd\") pod \"crc-debug-h4xzx\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:04 crc kubenswrapper[4688]: I1001 16:49:04.678656 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:05 crc kubenswrapper[4688]: I1001 16:49:05.188460 4688 generic.go:334] "Generic (PLEG): container finished" podID="ea295287-361f-4624-960e-ca4c4c9ec908" containerID="6deaf65f9a8181ec3fa18fbc4b5cd37a5c9f7f4c75af8dc1e186e619714886fa" exitCode=0 Oct 01 16:49:05 crc kubenswrapper[4688]: I1001 16:49:05.188576 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" event={"ID":"ea295287-361f-4624-960e-ca4c4c9ec908","Type":"ContainerDied","Data":"6deaf65f9a8181ec3fa18fbc4b5cd37a5c9f7f4c75af8dc1e186e619714886fa"} Oct 01 16:49:05 crc kubenswrapper[4688]: I1001 16:49:05.188751 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" event={"ID":"ea295287-361f-4624-960e-ca4c4c9ec908","Type":"ContainerStarted","Data":"2f54053d1c6f49ae26c7e63c3dc8e13488ea2e4dea66128f514fb216e1979e8f"} Oct 01 16:49:05 crc kubenswrapper[4688]: I1001 16:49:05.224450 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-h4xzx"] Oct 01 16:49:05 crc kubenswrapper[4688]: I1001 16:49:05.234279 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fbcdq/crc-debug-h4xzx"] Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.302938 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.381472 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:49:06 crc kubenswrapper[4688]: E1001 16:49:06.381881 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.448171 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zfsd\" (UniqueName: \"kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd\") pod \"ea295287-361f-4624-960e-ca4c4c9ec908\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.448363 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host\") pod \"ea295287-361f-4624-960e-ca4c4c9ec908\" (UID: \"ea295287-361f-4624-960e-ca4c4c9ec908\") " Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.448546 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host" (OuterVolumeSpecName: "host") pod "ea295287-361f-4624-960e-ca4c4c9ec908" (UID: "ea295287-361f-4624-960e-ca4c4c9ec908"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.448981 4688 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ea295287-361f-4624-960e-ca4c4c9ec908-host\") on node \"crc\" DevicePath \"\"" Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.462732 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd" (OuterVolumeSpecName: "kube-api-access-6zfsd") pod "ea295287-361f-4624-960e-ca4c4c9ec908" (UID: "ea295287-361f-4624-960e-ca4c4c9ec908"). InnerVolumeSpecName "kube-api-access-6zfsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:49:06 crc kubenswrapper[4688]: I1001 16:49:06.550329 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zfsd\" (UniqueName: \"kubernetes.io/projected/ea295287-361f-4624-960e-ca4c4c9ec908-kube-api-access-6zfsd\") on node \"crc\" DevicePath \"\"" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.183381 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.205550 4688 scope.go:117] "RemoveContainer" containerID="6deaf65f9a8181ec3fa18fbc4b5cd37a5c9f7f4c75af8dc1e186e619714886fa" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.205607 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/crc-debug-h4xzx" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.390294 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea295287-361f-4624-960e-ca4c4c9ec908" path="/var/lib/kubelet/pods/ea295287-361f-4624-960e-ca4c4c9ec908/volumes" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.447188 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.456863 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.457209 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.724712 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/util/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.740395 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/pull/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.750706 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_732f19e9db92c50483c3b3bb60cb9675cd299f69026bace77ce6e99233hxdmg_2f9c1b91-3db2-4b7e-8938-829c2c0d761f/extract/0.log" Oct 01 16:49:07 crc kubenswrapper[4688]: I1001 16:49:07.944027 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-65bw4_06f11b98-29b6-41f9-a9fa-771b2f9565ad/kube-rbac-proxy/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.101591 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-65bw4_06f11b98-29b6-41f9-a9fa-771b2f9565ad/manager/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.110791 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gbg2c_5c1fc770-71c1-4e90-b803-d0d020cf22dc/kube-rbac-proxy/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.258970 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-gbg2c_5c1fc770-71c1-4e90-b803-d0d020cf22dc/manager/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.367706 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-8pnfq_5fce237a-66be-49ef-90a8-cd0913daf883/manager/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.374674 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-8pnfq_5fce237a-66be-49ef-90a8-cd0913daf883/kube-rbac-proxy/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.526492 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-548l2_09bbe1ee-12b6-4342-b585-f5dfebdcfb79/kube-rbac-proxy/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.685470 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-548l2_09bbe1ee-12b6-4342-b585-f5dfebdcfb79/manager/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.785135 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-bh7tx_e983d733-09fb-4a40-902b-f180b58d5b7b/kube-rbac-proxy/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.854964 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-bh7tx_e983d733-09fb-4a40-902b-f180b58d5b7b/manager/0.log" Oct 01 16:49:08 crc kubenswrapper[4688]: I1001 16:49:08.955237 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5qm6b_44ef9a6f-0918-43c7-b153-abd8c7635131/kube-rbac-proxy/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.090495 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5qm6b_44ef9a6f-0918-43c7-b153-abd8c7635131/manager/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.116785 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-6ccdl_dc2fb11b-b04c-444c-8623-43a6b566cf6c/kube-rbac-proxy/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.298857 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-6ccdl_dc2fb11b-b04c-444c-8623-43a6b566cf6c/manager/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.431028 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-b8dfc_2fbdd013-920f-4e46-8519-3f8bbe24f25d/manager/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.474101 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-b8dfc_2fbdd013-920f-4e46-8519-3f8bbe24f25d/kube-rbac-proxy/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.639714 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-85595f4675-l62s8_57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c/kube-rbac-proxy/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.794083 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-85595f4675-l62s8_57be3a0d-f9b2-468c-b7f3-fdc1ea4f513c/manager/0.log" Oct 01 16:49:09 crc kubenswrapper[4688]: I1001 16:49:09.927922 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-mbsl5_45f48342-2b0d-4376-87ac-0c775c863326/kube-rbac-proxy/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.034445 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-mbsl5_45f48342-2b0d-4376-87ac-0c775c863326/manager/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.119022 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-x8lrd_8fee523c-d09e-422d-94ae-b85585856d27/kube-rbac-proxy/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.369941 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-x8lrd_8fee523c-d09e-422d-94ae-b85585856d27/manager/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.419057 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-h7xkj_6cfc7f51-a252-4b73-954c-b9bb611bf88e/kube-rbac-proxy/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.552828 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-h7xkj_6cfc7f51-a252-4b73-954c-b9bb611bf88e/manager/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.658055 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-9ppnb_3c33683d-9791-4d98-8d6e-0e58f48b4d94/kube-rbac-proxy/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.905382 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-9ppnb_3c33683d-9791-4d98-8d6e-0e58f48b4d94/manager/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.919907 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-rvtpg_8c13457f-36f6-49b5-9ac2-95298f76cd07/manager/0.log" Oct 01 16:49:10 crc kubenswrapper[4688]: I1001 16:49:10.933295 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-rvtpg_8c13457f-36f6-49b5-9ac2-95298f76cd07/kube-rbac-proxy/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.194360 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cs62rh_eb295639-6a81-4c5c-a02b-4101bf064788/kube-rbac-proxy/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.224860 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-77b9676b8cs62rh_eb295639-6a81-4c5c-a02b-4101bf064788/manager/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.415887 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c677785d7-6g6wn_7b66b48a-f1b3-4c89-a879-a662dce03612/kube-rbac-proxy/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.466450 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6445cc7789-t2zln_ded465e1-ef18-4321-9771-e3a3dd7a92e8/kube-rbac-proxy/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.660310 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6445cc7789-t2zln_ded465e1-ef18-4321-9771-e3a3dd7a92e8/operator/0.log" Oct 01 16:49:11 crc kubenswrapper[4688]: I1001 16:49:11.807001 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9cs6j_20aa9d1e-9bdb-45fb-8b83-ce5aac3f0010/registry-server/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.313001 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-tp7l5_fc3e31c2-228a-4ba3-aba1-113ef34d4917/kube-rbac-proxy/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.400178 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-tp7l5_fc3e31c2-228a-4ba3-aba1-113ef34d4917/manager/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.578451 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c677785d7-6g6wn_7b66b48a-f1b3-4c89-a879-a662dce03612/manager/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.604044 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g6725_7c87144a-f596-49d5-8344-b392d2bc4f51/kube-rbac-proxy/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.630785 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-g6725_7c87144a-f596-49d5-8344-b392d2bc4f51/manager/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.673345 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-w6mfn_fa800656-3342-44ac-9ec5-537382b5c67e/operator/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.850296 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-f4v99_604860e3-3d42-47b7-9387-c54faa3c8ec7/kube-rbac-proxy/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.862077 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-f4v99_604860e3-3d42-47b7-9387-c54faa3c8ec7/manager/0.log" Oct 01 16:49:12 crc kubenswrapper[4688]: I1001 16:49:12.946254 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-44nm7_a9b72717-b2f8-4340-81ed-88d2c3e0f2d6/kube-rbac-proxy/0.log" Oct 01 16:49:13 crc kubenswrapper[4688]: I1001 16:49:13.116053 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-44nm7_a9b72717-b2f8-4340-81ed-88d2c3e0f2d6/manager/0.log" Oct 01 16:49:13 crc kubenswrapper[4688]: I1001 16:49:13.150802 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-7blrb_733cc8e0-867c-4c7a-b535-c2493a08ed9f/manager/0.log" Oct 01 16:49:13 crc kubenswrapper[4688]: I1001 16:49:13.164615 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-7blrb_733cc8e0-867c-4c7a-b535-c2493a08ed9f/kube-rbac-proxy/0.log" Oct 01 16:49:13 crc kubenswrapper[4688]: I1001 16:49:13.314404 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-27q6x_52bede76-5e78-4a87-b1ed-4ec996866099/kube-rbac-proxy/0.log" Oct 01 16:49:13 crc kubenswrapper[4688]: I1001 16:49:13.364414 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-27q6x_52bede76-5e78-4a87-b1ed-4ec996866099/manager/0.log" Oct 01 16:49:17 crc kubenswrapper[4688]: I1001 16:49:17.381308 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:49:17 crc kubenswrapper[4688]: E1001 16:49:17.381954 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:49:31 crc kubenswrapper[4688]: I1001 16:49:31.388053 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:49:31 crc kubenswrapper[4688]: E1001 16:49:31.388835 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:49:32 crc kubenswrapper[4688]: I1001 16:49:32.430363 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9kj7x_dc8ba611-f5be-48f7-990f-f56d4f7a9539/control-plane-machine-set-operator/0.log" Oct 01 16:49:32 crc kubenswrapper[4688]: I1001 16:49:32.646023 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-szkjs_2f53e292-fdef-4716-b217-0c4fab2dd64c/kube-rbac-proxy/0.log" Oct 01 16:49:32 crc kubenswrapper[4688]: I1001 16:49:32.753016 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-szkjs_2f53e292-fdef-4716-b217-0c4fab2dd64c/machine-api-operator/0.log" Oct 01 16:49:42 crc kubenswrapper[4688]: I1001 16:49:42.381283 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:49:42 crc kubenswrapper[4688]: E1001 16:49:42.382122 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:49:45 crc kubenswrapper[4688]: I1001 16:49:45.410380 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ndwpx_6a722b12-5da2-4154-af09-c291e4530375/cert-manager-controller/0.log" Oct 01 16:49:45 crc kubenswrapper[4688]: I1001 16:49:45.604058 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-wb49t_4d4e2651-c2eb-4b0f-a3c1-aec4f0d9c95f/cert-manager-cainjector/0.log" Oct 01 16:49:45 crc kubenswrapper[4688]: I1001 16:49:45.661237 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-zxblg_6654a118-5437-4fb0-9c11-1ac4cf9dba79/cert-manager-webhook/0.log" Oct 01 16:49:54 crc kubenswrapper[4688]: I1001 16:49:54.382261 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:49:54 crc kubenswrapper[4688]: E1001 16:49:54.383208 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:49:58 crc kubenswrapper[4688]: I1001 16:49:58.602971 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-dwkqk_1be1c4ac-558b-4b81-97cf-feb86083dd5b/nmstate-console-plugin/0.log" Oct 01 16:49:58 crc kubenswrapper[4688]: I1001 16:49:58.933394 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2x5vh_8e78e6a0-4428-40b7-b819-6d798cf9983b/kube-rbac-proxy/0.log" Oct 01 16:49:58 crc kubenswrapper[4688]: I1001 16:49:58.945769 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mjqzm_7e9edca6-ee2b-4366-be0a-7c974e4865d7/nmstate-handler/0.log" Oct 01 16:49:59 crc kubenswrapper[4688]: I1001 16:49:59.034463 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-2x5vh_8e78e6a0-4428-40b7-b819-6d798cf9983b/nmstate-metrics/0.log" Oct 01 16:49:59 crc kubenswrapper[4688]: I1001 16:49:59.219061 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-mv2bb_bd754ddd-7d01-499a-9bfc-448c200a8c93/nmstate-operator/0.log" Oct 01 16:49:59 crc kubenswrapper[4688]: I1001 16:49:59.265366 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-x9vxk_67f45b7a-07fb-4cf3-b63d-832139f2bccd/nmstate-webhook/0.log" Oct 01 16:50:09 crc kubenswrapper[4688]: I1001 16:50:09.381662 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:50:09 crc kubenswrapper[4688]: E1001 16:50:09.382441 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.175929 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-st9jl_57161b05-13a8-415e-b9e9-ab575ff86007/kube-rbac-proxy/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.283696 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-st9jl_57161b05-13a8-415e-b9e9-ab575ff86007/controller/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.415315 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.582280 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.632147 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.639445 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.666139 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.817386 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.859078 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.900904 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:50:15 crc kubenswrapper[4688]: I1001 16:50:15.930271 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.049650 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-frr-files/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.071608 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-reloader/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.158127 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/controller/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.186177 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/cp-metrics/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.323215 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/frr-metrics/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.393339 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/kube-rbac-proxy/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.416255 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/frr/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.503678 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/kube-rbac-proxy-frr/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.545146 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pdvr5_d7efd23a-abc8-4bd4-9fd2-855fdbd9c2ac/reloader/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.606894 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-s7mkk_b047ab98-f421-44b0-9ee5-d2affd7804b8/frr-k8s-webhook-server/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.751806 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6887bf55bf-g25xn_550f3e3a-c6e4-4a2a-b0d3-8a995b6112a6/manager/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.852737 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-9b7c45d89-n5mmm_e8ccae4e-6cb9-40e8-b56b-4033f5d0f086/webhook-server/0.log" Oct 01 16:50:16 crc kubenswrapper[4688]: I1001 16:50:16.988655 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-j98l6_618070a7-23e1-4b85-a68e-85196f7838d9/kube-rbac-proxy/0.log" Oct 01 16:50:17 crc kubenswrapper[4688]: I1001 16:50:17.188186 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-j98l6_618070a7-23e1-4b85-a68e-85196f7838d9/speaker/0.log" Oct 01 16:50:21 crc kubenswrapper[4688]: I1001 16:50:21.388071 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:50:21 crc kubenswrapper[4688]: E1001 16:50:21.388931 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:50:29 crc kubenswrapper[4688]: I1001 16:50:29.973403 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.183445 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.212185 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.311389 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.405309 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/util/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.450501 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/pull/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.451115 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24p8jd_27aad031-d3e2-46e4-a73c-8a217125a386/extract/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.599600 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.780690 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.792604 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:50:30 crc kubenswrapper[4688]: I1001 16:50:30.798554 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.036156 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/util/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.057936 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/extract/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.082580 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc8tmc5_cac09066-7376-46af-805b-8d55c8d4045f/pull/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.232817 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.441118 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.444700 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.450216 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.603354 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-content/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.636172 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/extract-utilities/0.log" Oct 01 16:50:31 crc kubenswrapper[4688]: I1001 16:50:31.886309 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.061666 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-77ftd_25ac9f9e-1a94-4d95-bb20-20de73ca538a/registry-server/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.222300 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.234723 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.286510 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.401166 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-utilities/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.454907 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/extract-content/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.706011 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.888285 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:50:32 crc kubenswrapper[4688]: I1001 16:50:32.981946 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gpqvx_5df443ab-56dd-4124-963e-8e109914ea7e/registry-server/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.050471 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.052858 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.196725 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/util/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.207194 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/pull/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.250778 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96mgtcb_45408b58-1f2f-4f6b-a526-500f893245f0/extract/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.681499 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.877129 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.877903 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:50:33 crc kubenswrapper[4688]: I1001 16:50:33.889219 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.070830 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/util/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.081305 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/extract/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.096732 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kszs7_a368e405-b690-47e8-b2e7-ff9917147420/marketplace-operator/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.137275 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cfjqrv_6fa5d90a-0d97-4e6b-a5b7-88030e785f48/pull/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.284186 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.418223 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.431087 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.457555 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.623403 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-utilities/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.687250 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/extract-content/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.689248 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.769365 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m242n_092133e9-cde7-4be0-b5c2-e5de34a02727/registry-server/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.877095 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.916858 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:50:34 crc kubenswrapper[4688]: I1001 16:50:34.946227 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:50:35 crc kubenswrapper[4688]: I1001 16:50:35.092865 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-utilities/0.log" Oct 01 16:50:35 crc kubenswrapper[4688]: I1001 16:50:35.140820 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/extract-content/0.log" Oct 01 16:50:35 crc kubenswrapper[4688]: I1001 16:50:35.305134 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz9z6_b8cf2c87-9327-4dbd-b969-5a0d4000fd3f/registry-server/0.log" Oct 01 16:50:35 crc kubenswrapper[4688]: I1001 16:50:35.384380 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:50:35 crc kubenswrapper[4688]: E1001 16:50:35.384739 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:50:49 crc kubenswrapper[4688]: I1001 16:50:49.382577 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:50:49 crc kubenswrapper[4688]: E1001 16:50:49.383566 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:51:03 crc kubenswrapper[4688]: I1001 16:51:03.384190 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:51:03 crc kubenswrapper[4688]: E1001 16:51:03.384892 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:51:17 crc kubenswrapper[4688]: I1001 16:51:17.383169 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:51:17 crc kubenswrapper[4688]: E1001 16:51:17.383783 4688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6ndlj_openshift-machine-config-operator(69fd7067-ac62-426d-b3a3-dca2750d14db)\"" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" Oct 01 16:51:29 crc kubenswrapper[4688]: I1001 16:51:29.382157 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:51:30 crc kubenswrapper[4688]: I1001 16:51:30.490853 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"968ed772c564046bdc185c27e33824dba7083f8dfe024912082905d91f33acfd"} Oct 01 16:52:45 crc kubenswrapper[4688]: I1001 16:52:45.239177 4688 generic.go:334] "Generic (PLEG): container finished" podID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerID="646de5c380721cd02718ac8abfd76c4ffe1cdae078131c3fe91f5ea06a4d77e4" exitCode=0 Oct 01 16:52:45 crc kubenswrapper[4688]: I1001 16:52:45.239313 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" event={"ID":"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0","Type":"ContainerDied","Data":"646de5c380721cd02718ac8abfd76c4ffe1cdae078131c3fe91f5ea06a4d77e4"} Oct 01 16:52:45 crc kubenswrapper[4688]: I1001 16:52:45.240260 4688 scope.go:117] "RemoveContainer" containerID="646de5c380721cd02718ac8abfd76c4ffe1cdae078131c3fe91f5ea06a4d77e4" Oct 01 16:52:46 crc kubenswrapper[4688]: I1001 16:52:46.212389 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fbcdq_must-gather-xkt8s_d6e69463-ab7c-4ff9-9792-6119a6d7dfa0/gather/0.log" Oct 01 16:52:59 crc kubenswrapper[4688]: I1001 16:52:59.786199 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fbcdq/must-gather-xkt8s"] Oct 01 16:52:59 crc kubenswrapper[4688]: I1001 16:52:59.787853 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="copy" containerID="cri-o://c8fe70c70373ac8500bf84b0e92bf1f6cd02b6d6ed38353f5e857bdf9aaa18fe" gracePeriod=2 Oct 01 16:52:59 crc kubenswrapper[4688]: I1001 16:52:59.799432 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fbcdq/must-gather-xkt8s"] Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.397365 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fbcdq_must-gather-xkt8s_d6e69463-ab7c-4ff9-9792-6119a6d7dfa0/copy/0.log" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.398010 4688 generic.go:334] "Generic (PLEG): container finished" podID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerID="c8fe70c70373ac8500bf84b0e92bf1f6cd02b6d6ed38353f5e857bdf9aaa18fe" exitCode=143 Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.398060 4688 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="198a4da69446b6a0954c067a0e535fa19ba2f186be2bbf7b540a808ac6b0ef5a" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.481244 4688 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fbcdq_must-gather-xkt8s_d6e69463-ab7c-4ff9-9792-6119a6d7dfa0/copy/0.log" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.481721 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.541941 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fh4t\" (UniqueName: \"kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t\") pod \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.542152 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output\") pod \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\" (UID: \"d6e69463-ab7c-4ff9-9792-6119a6d7dfa0\") " Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.549185 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t" (OuterVolumeSpecName: "kube-api-access-8fh4t") pod "d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" (UID: "d6e69463-ab7c-4ff9-9792-6119a6d7dfa0"). InnerVolumeSpecName "kube-api-access-8fh4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.645472 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fh4t\" (UniqueName: \"kubernetes.io/projected/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-kube-api-access-8fh4t\") on node \"crc\" DevicePath \"\"" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.739940 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" (UID: "d6e69463-ab7c-4ff9-9792-6119a6d7dfa0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:53:00 crc kubenswrapper[4688]: I1001 16:53:00.747005 4688 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 16:53:01 crc kubenswrapper[4688]: I1001 16:53:01.393028 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" path="/var/lib/kubelet/pods/d6e69463-ab7c-4ff9-9792-6119a6d7dfa0/volumes" Oct 01 16:53:01 crc kubenswrapper[4688]: I1001 16:53:01.404693 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fbcdq/must-gather-xkt8s" Oct 01 16:53:03 crc kubenswrapper[4688]: I1001 16:53:03.551420 4688 scope.go:117] "RemoveContainer" containerID="c8fe70c70373ac8500bf84b0e92bf1f6cd02b6d6ed38353f5e857bdf9aaa18fe" Oct 01 16:53:03 crc kubenswrapper[4688]: I1001 16:53:03.589232 4688 scope.go:117] "RemoveContainer" containerID="04f959d4a9d42ca22a17d4f16c478707022ea230c2b77f8486417716f359f1c0" Oct 01 16:53:03 crc kubenswrapper[4688]: I1001 16:53:03.629365 4688 scope.go:117] "RemoveContainer" containerID="646de5c380721cd02718ac8abfd76c4ffe1cdae078131c3fe91f5ea06a4d77e4" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.111911 4688 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:28 crc kubenswrapper[4688]: E1001 16:53:28.114552 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea295287-361f-4624-960e-ca4c4c9ec908" containerName="container-00" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.114655 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea295287-361f-4624-960e-ca4c4c9ec908" containerName="container-00" Oct 01 16:53:28 crc kubenswrapper[4688]: E1001 16:53:28.114796 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="copy" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.114878 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="copy" Oct 01 16:53:28 crc kubenswrapper[4688]: E1001 16:53:28.114963 4688 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="gather" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.115042 4688 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="gather" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.115382 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea295287-361f-4624-960e-ca4c4c9ec908" containerName="container-00" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.115483 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="gather" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.115608 4688 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e69463-ab7c-4ff9-9792-6119a6d7dfa0" containerName="copy" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.117553 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.129577 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.223806 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.223893 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.224038 4688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kptfk\" (UniqueName: \"kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.355274 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.355343 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.355367 4688 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kptfk\" (UniqueName: \"kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.355887 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.356136 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.377945 4688 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kptfk\" (UniqueName: \"kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk\") pod \"certified-operators-sjcjv\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:28 crc kubenswrapper[4688]: I1001 16:53:28.450133 4688 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:29 crc kubenswrapper[4688]: I1001 16:53:29.032171 4688 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:29 crc kubenswrapper[4688]: I1001 16:53:29.709990 4688 generic.go:334] "Generic (PLEG): container finished" podID="eb825f0e-b6e4-4062-b37f-de9ad6e77f92" containerID="7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4" exitCode=0 Oct 01 16:53:29 crc kubenswrapper[4688]: I1001 16:53:29.710062 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerDied","Data":"7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4"} Oct 01 16:53:29 crc kubenswrapper[4688]: I1001 16:53:29.710105 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerStarted","Data":"a1e7ee9884d2bfb1a16f69361bb0aaf9040ad505b1753475fc33dc48dde1294e"} Oct 01 16:53:29 crc kubenswrapper[4688]: I1001 16:53:29.715863 4688 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 16:53:31 crc kubenswrapper[4688]: I1001 16:53:31.732322 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerStarted","Data":"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d"} Oct 01 16:53:33 crc kubenswrapper[4688]: I1001 16:53:33.751154 4688 generic.go:334] "Generic (PLEG): container finished" podID="eb825f0e-b6e4-4062-b37f-de9ad6e77f92" containerID="963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d" exitCode=0 Oct 01 16:53:33 crc kubenswrapper[4688]: I1001 16:53:33.751185 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerDied","Data":"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d"} Oct 01 16:53:34 crc kubenswrapper[4688]: I1001 16:53:34.765487 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerStarted","Data":"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a"} Oct 01 16:53:34 crc kubenswrapper[4688]: I1001 16:53:34.808753 4688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sjcjv" podStartSLOduration=2.016293308 podStartE2EDuration="6.808731168s" podCreationTimestamp="2025-10-01 16:53:28 +0000 UTC" firstStartedPulling="2025-10-01 16:53:29.71537003 +0000 UTC m=+3999.066010032" lastFinishedPulling="2025-10-01 16:53:34.50780793 +0000 UTC m=+4003.858447892" observedRunningTime="2025-10-01 16:53:34.800869924 +0000 UTC m=+4004.151509896" watchObservedRunningTime="2025-10-01 16:53:34.808731168 +0000 UTC m=+4004.159371120" Oct 01 16:53:38 crc kubenswrapper[4688]: I1001 16:53:38.451232 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:38 crc kubenswrapper[4688]: I1001 16:53:38.451836 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:38 crc kubenswrapper[4688]: I1001 16:53:38.502551 4688 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:48 crc kubenswrapper[4688]: I1001 16:53:48.498042 4688 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:48 crc kubenswrapper[4688]: I1001 16:53:48.549076 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:48 crc kubenswrapper[4688]: I1001 16:53:48.895326 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sjcjv" podUID="eb825f0e-b6e4-4062-b37f-de9ad6e77f92" containerName="registry-server" containerID="cri-o://b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a" gracePeriod=2 Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.466683 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.491147 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kptfk\" (UniqueName: \"kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk\") pod \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.491224 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities\") pod \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.491297 4688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content\") pod \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\" (UID: \"eb825f0e-b6e4-4062-b37f-de9ad6e77f92\") " Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.494091 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities" (OuterVolumeSpecName: "utilities") pod "eb825f0e-b6e4-4062-b37f-de9ad6e77f92" (UID: "eb825f0e-b6e4-4062-b37f-de9ad6e77f92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.497710 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk" (OuterVolumeSpecName: "kube-api-access-kptfk") pod "eb825f0e-b6e4-4062-b37f-de9ad6e77f92" (UID: "eb825f0e-b6e4-4062-b37f-de9ad6e77f92"). InnerVolumeSpecName "kube-api-access-kptfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.579870 4688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb825f0e-b6e4-4062-b37f-de9ad6e77f92" (UID: "eb825f0e-b6e4-4062-b37f-de9ad6e77f92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.593894 4688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kptfk\" (UniqueName: \"kubernetes.io/projected/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-kube-api-access-kptfk\") on node \"crc\" DevicePath \"\"" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.593929 4688 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.593941 4688 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb825f0e-b6e4-4062-b37f-de9ad6e77f92-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.906934 4688 generic.go:334] "Generic (PLEG): container finished" podID="eb825f0e-b6e4-4062-b37f-de9ad6e77f92" containerID="b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a" exitCode=0 Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.907027 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerDied","Data":"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a"} Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.907180 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjcjv" event={"ID":"eb825f0e-b6e4-4062-b37f-de9ad6e77f92","Type":"ContainerDied","Data":"a1e7ee9884d2bfb1a16f69361bb0aaf9040ad505b1753475fc33dc48dde1294e"} Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.907201 4688 scope.go:117] "RemoveContainer" containerID="b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.907043 4688 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjcjv" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.938730 4688 scope.go:117] "RemoveContainer" containerID="963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d" Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.945885 4688 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.959779 4688 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sjcjv"] Oct 01 16:53:49 crc kubenswrapper[4688]: I1001 16:53:49.980417 4688 scope.go:117] "RemoveContainer" containerID="7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.025001 4688 scope.go:117] "RemoveContainer" containerID="b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a" Oct 01 16:53:50 crc kubenswrapper[4688]: E1001 16:53:50.025745 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a\": container with ID starting with b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a not found: ID does not exist" containerID="b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.025787 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a"} err="failed to get container status \"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a\": rpc error: code = NotFound desc = could not find container \"b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a\": container with ID starting with b466475e35ad99bc8553b38f95f22e516730520eb764c6209e1b8417cff2917a not found: ID does not exist" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.025814 4688 scope.go:117] "RemoveContainer" containerID="963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d" Oct 01 16:53:50 crc kubenswrapper[4688]: E1001 16:53:50.026057 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d\": container with ID starting with 963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d not found: ID does not exist" containerID="963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.026073 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d"} err="failed to get container status \"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d\": rpc error: code = NotFound desc = could not find container \"963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d\": container with ID starting with 963888e8578cafb8dba480a20eeef98e31cb31a289045be670bdc8bc90581f2d not found: ID does not exist" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.026084 4688 scope.go:117] "RemoveContainer" containerID="7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4" Oct 01 16:53:50 crc kubenswrapper[4688]: E1001 16:53:50.026263 4688 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4\": container with ID starting with 7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4 not found: ID does not exist" containerID="7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4" Oct 01 16:53:50 crc kubenswrapper[4688]: I1001 16:53:50.026280 4688 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4"} err="failed to get container status \"7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4\": rpc error: code = NotFound desc = could not find container \"7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4\": container with ID starting with 7beb2ec7320dddebd4e399d64386d58ab0c0872a12c23fc5677f19b6fa78c4b4 not found: ID does not exist" Oct 01 16:53:51 crc kubenswrapper[4688]: I1001 16:53:51.397123 4688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb825f0e-b6e4-4062-b37f-de9ad6e77f92" path="/var/lib/kubelet/pods/eb825f0e-b6e4-4062-b37f-de9ad6e77f92/volumes" Oct 01 16:53:55 crc kubenswrapper[4688]: I1001 16:53:55.052981 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:53:55 crc kubenswrapper[4688]: I1001 16:53:55.053483 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:54:25 crc kubenswrapper[4688]: I1001 16:54:25.053396 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:54:25 crc kubenswrapper[4688]: I1001 16:54:25.054090 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.052638 4688 patch_prober.go:28] interesting pod/machine-config-daemon-6ndlj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.053253 4688 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.053309 4688 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.054105 4688 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"968ed772c564046bdc185c27e33824dba7083f8dfe024912082905d91f33acfd"} pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.054173 4688 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" podUID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerName="machine-config-daemon" containerID="cri-o://968ed772c564046bdc185c27e33824dba7083f8dfe024912082905d91f33acfd" gracePeriod=600 Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.613335 4688 generic.go:334] "Generic (PLEG): container finished" podID="69fd7067-ac62-426d-b3a3-dca2750d14db" containerID="968ed772c564046bdc185c27e33824dba7083f8dfe024912082905d91f33acfd" exitCode=0 Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.613682 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerDied","Data":"968ed772c564046bdc185c27e33824dba7083f8dfe024912082905d91f33acfd"} Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.613719 4688 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6ndlj" event={"ID":"69fd7067-ac62-426d-b3a3-dca2750d14db","Type":"ContainerStarted","Data":"132811a18ec8c91a0aa5d37ebc0baa2053a14c7238f848298002488b3b9af252"} Oct 01 16:54:55 crc kubenswrapper[4688]: I1001 16:54:55.613753 4688 scope.go:117] "RemoveContainer" containerID="6176a2d7d9bf07039df55be0f0cea1a4049132ea682ab74726cc5e797be99137" Oct 01 16:55:03 crc kubenswrapper[4688]: I1001 16:55:03.781598 4688 scope.go:117] "RemoveContainer" containerID="32a0d78a9e94b1c5a7928ed23dfe36a1afaddf6ee6e73dc46d17c7d49eaf48e3" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067256374024464 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067256374017401 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067245733016521 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067245733015471 5ustar corecore